TFX Standard Components Walkthrough Reviews
4445 reviews
Could not get python imports to work
Abdul Z. · Reviewed over 2 years ago
it still doesn't work! Now Jupiterlab won't open.
Paul M. · Reviewed over 2 years ago
The lab is not compatible with python 3 anymore... the lab can not be completed
Tamás P. · Reviewed over 2 years ago
Dariusz Ł. · Reviewed over 2 years ago
Codes not running or compatible to latest versions
suvojyoti c. · Reviewed over 2 years ago
Tanai A. · Reviewed over 2 years ago
Error on `pip install tensorflow_model_analysis` gcc: build/src.linux-x86_64-3.10/numpy/core/src/multiarray/scalartypes.c numpy/core/src/multiarray/scalartypes.c.src: In function ‘float_arrtype_hash’: numpy/core/src/multiarray/scalartypes.c.src:2967:27: error: incompatible type for argument 1 of ‘_Py_HashDouble’ 2967 | return _Py_HashDouble((double) PyArrayScalar_VAL(obj, @name@)); In file included from /opt/conda/include/python3.10/Python.h:77, from numpy/core/src/multiarray/scalartypes.c.src:3: /opt/conda/include/python3.10/pyhash.h:10:38: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘double’ 10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double); | ^~~~~~~~~~ numpy/core/src/multiarray/scalartypes.c.src:2967:12: error: too few arguments to function ‘_Py_HashDouble’ 2967 | return _Py_HashDouble((double) PyArrayScalar_VAL(obj, @name@)); | ^~~~~~~~~~~~~~ In file included from /opt/conda/include/python3.10/Python.h:77, from numpy/core/src/multiarray/scalartypes.c.src:3: /opt/conda/include/python3.10/pyhash.h:10:23: note: declared here 10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double); | ^~~~~~~~~~~~~~ numpy/core/src/multiarray/scalartypes.c.src: In function ‘cfloat_arrtype_hash’: numpy/core/src/multiarray/scalartypes.c.src:2975:31: error: incompatible type for argument 1 of ‘_Py_HashDouble’ 2975 | hashreal = _Py_HashDouble((double) | ^~~~~~~~ | | | double 2976 | PyArrayScalar_VAL(obj, C@name@).real); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ In file included from /opt/conda/include/python3.10/Python.h:77, from numpy/core/src/multiarray/scalartypes.c.src:3: /opt/conda/include/python3.10/pyhash.h:10:38: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘double’ 10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double); | ^~~~~~~~~~ numpy/core/src/multiarray/scalartypes.c.src:2975:16: error: too few arguments to function ‘_Py_HashDouble’ 2975 | hashreal = _Py_HashDouble((double) | ^~~~~~~~~~~~~~ In file included from /opt/conda/include/python3.10/Python.h:77, from numpy/core/src/multiarray/scalartypes.c.src:3: /opt/conda/include/python3.10/pyhash.h:10:23: note: declared here 10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double); | ^~~~~~~~~~~~~~ numpy/core/src/multiarray/scalartypes.c.src:2981:31: error: incompatible type for argument 1 of ‘_Py_HashDouble’ 2981 | hashimag = _Py_HashDouble((double) | ^~~~~~~~ | | | double 2982 | PyArrayScalar_VAL(obj, C@name@).imag); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ In file included from /opt/conda/include/python3.10/Python.h:77, from numpy/core/src/multiarray/scalartypes.c.src:3: /opt/conda/include/python3.10/pyhash.h:10:38: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘double’ 10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double); | ^~~~~~~~~~ numpy/core/src/multiarray/scalartypes.c.src:2981:16: error: too few arguments to function ‘_Py_HashDouble’ 2981 | hashimag = _Py_HashDouble((double) | ^~~~~~~~~~~~~~ In file included from /opt/conda/include/python3.10/Python.h:77, from numpy/core/src/multiarray/scalartypes.c.src:3: /opt/conda/include/python3.10/pyhash.h:10:23: note: declared here 10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double); | ^~~~~~~~~~~~~~ numpy/core/src/multiarray/scalartypes.c.src: In function ‘longdouble_arrtype_hash’: numpy/core/src/multiarray/scalartypes.c.src:2967:27: error: incompatible type for argument 1 of ‘_Py_HashDouble’ 2967 | return _Py_HashDouble((double) PyArrayScalar_VAL(obj, @name@)); In file included from /opt/conda/include/python3.10/Python.h:77, from numpy/core/src/multiarray/scalartypes.c.src:3: /opt/conda/include/python3.10/pyhash.h:10:38: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘double’ 10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double); | ^~~~~~~~~~ numpy/core/src/multiarray/scalartypes.c.src:2967:12: error: too few arguments to function ‘_Py_HashDouble’ 2967 | return _Py_HashDouble((double) PyArrayScalar_VAL(obj, @name@)); | ^~~~~~~~~~~~~~ In file included from /opt/conda/include/python3.10/Python.h:77, from numpy/core/src/multiarray/scalartypes.c.src:3: /opt/conda/include/python3.10/pyhash.h:10:23: note: declared here 10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double); | ^~~~~~~~~~~~~~ numpy/core/src/multiarray/scalartypes.c.src: In function ‘clongdouble_arrtype_hash’: numpy/core/src/multiarray/scalartypes.c.src:2975:31: error: incompatible type for argument 1 of ‘_Py_HashDouble’ 2975 | hashreal = _Py_HashDouble((double) | ^~~~~~~~ | | | double 2976 | PyArrayScalar_VAL(obj, C@name@).real); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ In file included from /opt/conda/include/python3.10/Python.h:77, from numpy/core/src/multiarray/scalartypes.c.src:3: /opt/conda/include/python3.10/pyhash.h:10:38: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘double’ 10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double); | ^~~~~~~~~~ numpy/core/src/multiarray/scalartypes.c.src:2975:16: error: too few arguments to function ‘_Py_HashDouble’ 2975 | hashreal = _Py_HashDouble((double) | ^~~~~~~~~~~~~~ In file included from /opt/conda/include/python3.10/Python.h:77, from numpy/core/src/multiarray/scalartypes.c.src:3: /opt/conda/include/python3.10/pyhash.h:10:23: note: declared here 10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double); | ^~~~~~~~~~~~~~ numpy/core/src/multiarray/scalartypes.c.src:2981:31: error: incompatible type for argument 1 of ‘_Py_HashDouble’ 2981 | hashimag = _Py_HashDouble((double) | ^~~~~~~~ | | | double 2982 | PyArrayScalar_VAL(obj, C@name@).imag); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ In file included from /opt/conda/include/python3.10/Python.h:77, from numpy/core/src/multiarray/scalartypes.c.src:3: /opt/conda/include/python3.10/pyhash.h:10:38: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘double’ 10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double); | ^~~~~~~~~~ numpy/core/src/multiarray/scalartypes.c.src:2981:16: error: too few arguments to function ‘_Py_HashDouble’ 2981 | hashimag = _Py_HashDouble((double) | ^~~~~~~~~~~~~~ In file included from /opt/conda/include/python3.10/Python.h:77, from numpy/core/src/multiarray/scalartypes.c.src:3: /opt/conda/include/python3.10/pyhash.h:10:23: note: declared here 10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double); | ^~~~~~~~~~~~~~ numpy/core/src/multiarray/scalartypes.c.src: In function ‘half_arrtype_hash’: numpy/core/src/multiarray/scalartypes.c.src:2997:27: error: incompatible type for argument 1 of ‘_Py_HashDouble’ 2997 | return _Py_HashDouble(npy_half_to_double(PyArrayScalar_VAL(obj, Half))); | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ | | | double In file included from /opt/conda/include/python3.10/Python.h:77, from numpy/core/src/multiarray/scalartypes.c.src:3: /opt/conda/include/python3.10/pyhash.h:10:38: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘double’ 10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double); | ^~~~~~~~~~ numpy/core/src/multiarray/scalartypes.c.src:2997:12: error: too few arguments to function ‘_Py_HashDouble’ 2997 | return _Py_HashDouble(npy_half_to_double(PyArrayScalar_VAL(obj, Half))); | ^~~~~~~~~~~~~~ In file included from /opt/conda/include/python3.10/Python.h:77, from numpy/core/src/multiarray/scalartypes.c.src:3: /opt/conda/include/python3.10/pyhash.h:10:23: note: declared here 10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double); | ^~~~~~~~~~~~~~ numpy/core/src/multiarray/scalartypes.c.src: In function ‘longdouble_arrtype_hash’: numpy/core/src/multiarray/scalartypes.c.src:2968:1: warning: control reaches end of non-void function [-Wreturn-type] 2968 | } | ^ numpy/core/src/multiarray/scalartypes.c.src: In function ‘float_arrtype_hash’: numpy/core/src/multiarray/scalartypes.c.src:2968:1: warning: control reaches end of non-void function [-Wreturn-type] 2968 | } | ^ numpy/core/src/multiarray/scalartypes.c.src: In function ‘half_arrtype_hash’: numpy/core/src/multiarray/scalartypes.c.src:2998:1: warning: control reaches end of non-void function [-Wreturn-type] 2998 | } | ^ gcc: numpy/core/src/multiarray/scalarapi.c gcc: numpy/core/src/multiarray/vdot.c gcc: numpy/core/src/multiarray/nditer_pywrap.c gcc: numpy/core/src/umath/umathmodule.c gcc: build/src.linux-x86_64-3.10/numpy/core/src/umath/clip.c gcc: numpy/core/src/umath/reduction.c gcc: build/src.linux-x86_64-3.10/numpy/core/src/umath/loops.c gcc: numpy/core/src/umath/override.c gcc: numpy/core/src/umath/ufunc_object.c gcc: numpy/core/src/npymath/npy_math.c gcc: build/src.linux-x86_64-3.10/numpy/core/src/npymath/ieee754.c gcc: build/src.linux-x86_64-3.10/numpy/core/src/npymath/npy_math_complex.c gcc: numpy/core/src/npymath/halffloat.c gcc: numpy/core/src/common/array_assign.c gcc: numpy/core/src/common/mem_overlap.c gcc: numpy/core/src/umath/extobj.c gcc: numpy/core/src/common/npy_longdouble.c gcc: build/src.linux-x86_64-3.10/numpy/core/src/umath/scalarmath.c gcc: numpy/core/src/common/ucsnarrow.c gcc: numpy/core/src/common/ufunc_override.c gcc: numpy/core/src/common/numpyos.c gcc: build/src.linux-x86_64-3.10/numpy/core/src/common/npy_cpu_features.c gcc: numpy/core/src/umath/ufunc_type_resolution.c gcc: numpy/core/src/multiarray/mapping.c gcc: numpy/core/src/multiarray/methods.c gcc: build/src.linux-x86_64-3.10/numpy/core/src/umath/matmul.c error: Command "gcc -pthread -B /opt/conda/compiler_compat -Wno-unused-result -Wsign-compare -DNDEBUG -fwrapv -O2 -Wall -fPIC -O2 -isystem /opt/conda/include -fPIC -O2 -isystem /opt/conda/include -fPIC -DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Ibuild/src.linux-x86_64-3.10/numpy/core/src/umath -Ibuild/src.linux-x86_64-3.10/numpy/core/src/npymath -Ibuild/src.linux-x86_64-3.10/numpy/core/src/common -Inumpy/core/include -Ibuild/src.linux-x86_64-3.10/numpy/core/include/numpy -Inumpy/core/src/common -Inumpy/core/src -Inumpy/core -Inumpy/core/src/npymath -Inumpy/core/src/multiarray -Inumpy/core/src/umath -Inumpy/core/src/npysort -I/opt/conda/include/python3.10 -Ibuild/src.linux-x86_64-3.10/numpy/core/src/common -Ibuild/src.linux-x86_64-3.10/numpy/core/src/npymath -c build/src.linux-x86_64-3.10/numpy/core/src/multiarray/scalartypes.c -o build/temp.linux-x86_64-3.10/build/src.linux-x86_64-3.10/numpy/core/src/multiarray/scalartypes.o -MMD -MF build/temp.linux-x86_64-3.10/build/src.linux-x86_64-3.10/numpy/core/src/multiarray/scalartypes.o.d" failed with exit status 1 [end of output] note: This error originates from a subprocess, and is likely not a problem with pip. ERROR: Failed building wheel for numpy Failed to build numpy ERROR: Could not build wheels for numpy, which is required to install pyproject.toml-based projects [end of output] note: This error originates from a subprocess, and is likely not a problem with pip.
ChengCheng T. · Reviewed over 2 years ago
lab does not work with current notebook settings
Guendalina C. · Reviewed over 2 years ago
Tamás P. · Reviewed over 2 years ago
Faulty instructions, there is no AI Platform, there is no create instance on Vertex AI Pipelines page. Tensorflow data validation library is not installed and installing permission is denied.
Arslan A. · Reviewed over 2 years ago
Arvindh R. · Reviewed over 2 years ago
--------------------------------------------------------------------------- ModuleNotFoundError Traceback (most recent call last) Cell In[1], line 7 4 import time 6 import tensorflow as tf ----> 7 import tensorflow_data_validation as tfdv
Skandarajan R. · Reviewed over 2 years ago
Carl M. · Reviewed over 2 years ago
please update the lab for latest version of tensorflow
Arvindh R. · Reviewed over 2 years ago
Federico F. · Reviewed over 2 years ago
Carl M. · Reviewed over 2 years ago
Mahdieh K. · Reviewed over 2 years ago
it still doesn't work! After 45 minutes received this error: Error encountered while creating a cluster. Please try again or select an existing cluster. Failure Reason: Error while creating the new cluster - Error: Cluster not in running state
Paul M. · Reviewed over 2 years ago
Eduardo J. · Reviewed over 2 years ago
I coul not create any pipeline it gave me an error. I spent 2 hours waiting for the cluster to be created and no success
Niloofar M. · Reviewed over 2 years ago
Marco M. · Reviewed over 2 years ago
Aleksei K. · Reviewed over 2 years ago
Doesn't work the deploy of kubeflow pipelines. Creating "custer-1" in zone "us-central1-a" doesm't finish
Ivan R. · Reviewed over 2 years ago
Ivan R. · Reviewed over 2 years ago
Anas K. · Reviewed over 2 years ago
We do not ensure the published reviews originate from consumers who have purchased or used the products. Reviews are not verified by Google.