TFX Standard Components Walkthrough Ulasan

4445 ulasan

Could not get python imports to work

Abdul Z. · Diulas lebih dari 2 tahun lalu

it still doesn't work! Now Jupiterlab won't open.

Paul M. · Diulas lebih dari 2 tahun lalu

The lab is not compatible with python 3 anymore... the lab can not be completed

Tamás P. · Diulas lebih dari 2 tahun lalu

Dariusz Ł. · Diulas lebih dari 2 tahun lalu

Codes not running or compatible to latest versions

suvojyoti c. · Diulas lebih dari 2 tahun lalu

Tanai A. · Diulas lebih dari 2 tahun lalu

Error on `pip install tensorflow_model_analysis` gcc: build/src.linux-x86_64-3.10/numpy/core/src/multiarray/scalartypes.c numpy/core/src/multiarray/scalartypes.c.src: In function ‘float_arrtype_hash’: numpy/core/src/multiarray/scalartypes.c.src:2967:27: error: incompatible type for argument 1 of ‘_Py_HashDouble’ 2967 | return _Py_HashDouble((double) PyArrayScalar_VAL(obj, @name@)); In file included from /opt/conda/include/python3.10/Python.h:77, from numpy/core/src/multiarray/scalartypes.c.src:3: /opt/conda/include/python3.10/pyhash.h:10:38: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘double’ 10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double); | ^~~~~~~~~~ numpy/core/src/multiarray/scalartypes.c.src:2967:12: error: too few arguments to function ‘_Py_HashDouble’ 2967 | return _Py_HashDouble((double) PyArrayScalar_VAL(obj, @name@)); | ^~~~~~~~~~~~~~ In file included from /opt/conda/include/python3.10/Python.h:77, from numpy/core/src/multiarray/scalartypes.c.src:3: /opt/conda/include/python3.10/pyhash.h:10:23: note: declared here 10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double); | ^~~~~~~~~~~~~~ numpy/core/src/multiarray/scalartypes.c.src: In function ‘cfloat_arrtype_hash’: numpy/core/src/multiarray/scalartypes.c.src:2975:31: error: incompatible type for argument 1 of ‘_Py_HashDouble’ 2975 | hashreal = _Py_HashDouble((double) | ^~~~~~~~ | | | double 2976 | PyArrayScalar_VAL(obj, C@name@).real); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ In file included from /opt/conda/include/python3.10/Python.h:77, from numpy/core/src/multiarray/scalartypes.c.src:3: /opt/conda/include/python3.10/pyhash.h:10:38: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘double’ 10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double); | ^~~~~~~~~~ numpy/core/src/multiarray/scalartypes.c.src:2975:16: error: too few arguments to function ‘_Py_HashDouble’ 2975 | hashreal = _Py_HashDouble((double) | ^~~~~~~~~~~~~~ In file included from /opt/conda/include/python3.10/Python.h:77, from numpy/core/src/multiarray/scalartypes.c.src:3: /opt/conda/include/python3.10/pyhash.h:10:23: note: declared here 10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double); | ^~~~~~~~~~~~~~ numpy/core/src/multiarray/scalartypes.c.src:2981:31: error: incompatible type for argument 1 of ‘_Py_HashDouble’ 2981 | hashimag = _Py_HashDouble((double) | ^~~~~~~~ | | | double 2982 | PyArrayScalar_VAL(obj, C@name@).imag); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ In file included from /opt/conda/include/python3.10/Python.h:77, from numpy/core/src/multiarray/scalartypes.c.src:3: /opt/conda/include/python3.10/pyhash.h:10:38: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘double’ 10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double); | ^~~~~~~~~~ numpy/core/src/multiarray/scalartypes.c.src:2981:16: error: too few arguments to function ‘_Py_HashDouble’ 2981 | hashimag = _Py_HashDouble((double) | ^~~~~~~~~~~~~~ In file included from /opt/conda/include/python3.10/Python.h:77, from numpy/core/src/multiarray/scalartypes.c.src:3: /opt/conda/include/python3.10/pyhash.h:10:23: note: declared here 10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double); | ^~~~~~~~~~~~~~ numpy/core/src/multiarray/scalartypes.c.src: In function ‘longdouble_arrtype_hash’: numpy/core/src/multiarray/scalartypes.c.src:2967:27: error: incompatible type for argument 1 of ‘_Py_HashDouble’ 2967 | return _Py_HashDouble((double) PyArrayScalar_VAL(obj, @name@)); In file included from /opt/conda/include/python3.10/Python.h:77, from numpy/core/src/multiarray/scalartypes.c.src:3: /opt/conda/include/python3.10/pyhash.h:10:38: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘double’ 10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double); | ^~~~~~~~~~ numpy/core/src/multiarray/scalartypes.c.src:2967:12: error: too few arguments to function ‘_Py_HashDouble’ 2967 | return _Py_HashDouble((double) PyArrayScalar_VAL(obj, @name@)); | ^~~~~~~~~~~~~~ In file included from /opt/conda/include/python3.10/Python.h:77, from numpy/core/src/multiarray/scalartypes.c.src:3: /opt/conda/include/python3.10/pyhash.h:10:23: note: declared here 10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double); | ^~~~~~~~~~~~~~ numpy/core/src/multiarray/scalartypes.c.src: In function ‘clongdouble_arrtype_hash’: numpy/core/src/multiarray/scalartypes.c.src:2975:31: error: incompatible type for argument 1 of ‘_Py_HashDouble’ 2975 | hashreal = _Py_HashDouble((double) | ^~~~~~~~ | | | double 2976 | PyArrayScalar_VAL(obj, C@name@).real); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ In file included from /opt/conda/include/python3.10/Python.h:77, from numpy/core/src/multiarray/scalartypes.c.src:3: /opt/conda/include/python3.10/pyhash.h:10:38: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘double’ 10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double); | ^~~~~~~~~~ numpy/core/src/multiarray/scalartypes.c.src:2975:16: error: too few arguments to function ‘_Py_HashDouble’ 2975 | hashreal = _Py_HashDouble((double) | ^~~~~~~~~~~~~~ In file included from /opt/conda/include/python3.10/Python.h:77, from numpy/core/src/multiarray/scalartypes.c.src:3: /opt/conda/include/python3.10/pyhash.h:10:23: note: declared here 10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double); | ^~~~~~~~~~~~~~ numpy/core/src/multiarray/scalartypes.c.src:2981:31: error: incompatible type for argument 1 of ‘_Py_HashDouble’ 2981 | hashimag = _Py_HashDouble((double) | ^~~~~~~~ | | | double 2982 | PyArrayScalar_VAL(obj, C@name@).imag); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ In file included from /opt/conda/include/python3.10/Python.h:77, from numpy/core/src/multiarray/scalartypes.c.src:3: /opt/conda/include/python3.10/pyhash.h:10:38: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘double’ 10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double); | ^~~~~~~~~~ numpy/core/src/multiarray/scalartypes.c.src:2981:16: error: too few arguments to function ‘_Py_HashDouble’ 2981 | hashimag = _Py_HashDouble((double) | ^~~~~~~~~~~~~~ In file included from /opt/conda/include/python3.10/Python.h:77, from numpy/core/src/multiarray/scalartypes.c.src:3: /opt/conda/include/python3.10/pyhash.h:10:23: note: declared here 10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double); | ^~~~~~~~~~~~~~ numpy/core/src/multiarray/scalartypes.c.src: In function ‘half_arrtype_hash’: numpy/core/src/multiarray/scalartypes.c.src:2997:27: error: incompatible type for argument 1 of ‘_Py_HashDouble’ 2997 | return _Py_HashDouble(npy_half_to_double(PyArrayScalar_VAL(obj, Half))); | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ | | | double In file included from /opt/conda/include/python3.10/Python.h:77, from numpy/core/src/multiarray/scalartypes.c.src:3: /opt/conda/include/python3.10/pyhash.h:10:38: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘double’ 10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double); | ^~~~~~~~~~ numpy/core/src/multiarray/scalartypes.c.src:2997:12: error: too few arguments to function ‘_Py_HashDouble’ 2997 | return _Py_HashDouble(npy_half_to_double(PyArrayScalar_VAL(obj, Half))); | ^~~~~~~~~~~~~~ In file included from /opt/conda/include/python3.10/Python.h:77, from numpy/core/src/multiarray/scalartypes.c.src:3: /opt/conda/include/python3.10/pyhash.h:10:23: note: declared here 10 | PyAPI_FUNC(Py_hash_t) _Py_HashDouble(PyObject *, double); | ^~~~~~~~~~~~~~ numpy/core/src/multiarray/scalartypes.c.src: In function ‘longdouble_arrtype_hash’: numpy/core/src/multiarray/scalartypes.c.src:2968:1: warning: control reaches end of non-void function [-Wreturn-type] 2968 | } | ^ numpy/core/src/multiarray/scalartypes.c.src: In function ‘float_arrtype_hash’: numpy/core/src/multiarray/scalartypes.c.src:2968:1: warning: control reaches end of non-void function [-Wreturn-type] 2968 | } | ^ numpy/core/src/multiarray/scalartypes.c.src: In function ‘half_arrtype_hash’: numpy/core/src/multiarray/scalartypes.c.src:2998:1: warning: control reaches end of non-void function [-Wreturn-type] 2998 | } | ^ gcc: numpy/core/src/multiarray/scalarapi.c gcc: numpy/core/src/multiarray/vdot.c gcc: numpy/core/src/multiarray/nditer_pywrap.c gcc: numpy/core/src/umath/umathmodule.c gcc: build/src.linux-x86_64-3.10/numpy/core/src/umath/clip.c gcc: numpy/core/src/umath/reduction.c gcc: build/src.linux-x86_64-3.10/numpy/core/src/umath/loops.c gcc: numpy/core/src/umath/override.c gcc: numpy/core/src/umath/ufunc_object.c gcc: numpy/core/src/npymath/npy_math.c gcc: build/src.linux-x86_64-3.10/numpy/core/src/npymath/ieee754.c gcc: build/src.linux-x86_64-3.10/numpy/core/src/npymath/npy_math_complex.c gcc: numpy/core/src/npymath/halffloat.c gcc: numpy/core/src/common/array_assign.c gcc: numpy/core/src/common/mem_overlap.c gcc: numpy/core/src/umath/extobj.c gcc: numpy/core/src/common/npy_longdouble.c gcc: build/src.linux-x86_64-3.10/numpy/core/src/umath/scalarmath.c gcc: numpy/core/src/common/ucsnarrow.c gcc: numpy/core/src/common/ufunc_override.c gcc: numpy/core/src/common/numpyos.c gcc: build/src.linux-x86_64-3.10/numpy/core/src/common/npy_cpu_features.c gcc: numpy/core/src/umath/ufunc_type_resolution.c gcc: numpy/core/src/multiarray/mapping.c gcc: numpy/core/src/multiarray/methods.c gcc: build/src.linux-x86_64-3.10/numpy/core/src/umath/matmul.c error: Command "gcc -pthread -B /opt/conda/compiler_compat -Wno-unused-result -Wsign-compare -DNDEBUG -fwrapv -O2 -Wall -fPIC -O2 -isystem /opt/conda/include -fPIC -O2 -isystem /opt/conda/include -fPIC -DNPY_INTERNAL_BUILD=1 -DHAVE_NPY_CONFIG_H=1 -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -D_LARGEFILE64_SOURCE=1 -Ibuild/src.linux-x86_64-3.10/numpy/core/src/umath -Ibuild/src.linux-x86_64-3.10/numpy/core/src/npymath -Ibuild/src.linux-x86_64-3.10/numpy/core/src/common -Inumpy/core/include -Ibuild/src.linux-x86_64-3.10/numpy/core/include/numpy -Inumpy/core/src/common -Inumpy/core/src -Inumpy/core -Inumpy/core/src/npymath -Inumpy/core/src/multiarray -Inumpy/core/src/umath -Inumpy/core/src/npysort -I/opt/conda/include/python3.10 -Ibuild/src.linux-x86_64-3.10/numpy/core/src/common -Ibuild/src.linux-x86_64-3.10/numpy/core/src/npymath -c build/src.linux-x86_64-3.10/numpy/core/src/multiarray/scalartypes.c -o build/temp.linux-x86_64-3.10/build/src.linux-x86_64-3.10/numpy/core/src/multiarray/scalartypes.o -MMD -MF build/temp.linux-x86_64-3.10/build/src.linux-x86_64-3.10/numpy/core/src/multiarray/scalartypes.o.d" failed with exit status 1 [end of output] note: This error originates from a subprocess, and is likely not a problem with pip. ERROR: Failed building wheel for numpy Failed to build numpy ERROR: Could not build wheels for numpy, which is required to install pyproject.toml-based projects [end of output] note: This error originates from a subprocess, and is likely not a problem with pip.

ChengCheng T. · Diulas lebih dari 2 tahun lalu

lab does not work with current notebook settings

Guendalina C. · Diulas lebih dari 2 tahun lalu

Tamás P. · Diulas lebih dari 2 tahun lalu

Faulty instructions, there is no AI Platform, there is no create instance on Vertex AI Pipelines page. Tensorflow data validation library is not installed and installing permission is denied.

Arslan A. · Diulas lebih dari 2 tahun lalu

Arvindh R. · Diulas lebih dari 2 tahun lalu

--------------------------------------------------------------------------- ModuleNotFoundError Traceback (most recent call last) Cell In[1], line 7 4 import time 6 import tensorflow as tf ----> 7 import tensorflow_data_validation as tfdv

Skandarajan R. · Diulas lebih dari 2 tahun lalu

Carl M. · Diulas lebih dari 2 tahun lalu

please update the lab for latest version of tensorflow

Arvindh R. · Diulas lebih dari 2 tahun lalu

Federico F. · Diulas lebih dari 2 tahun lalu

Carl M. · Diulas lebih dari 2 tahun lalu

Mahdieh K. · Diulas lebih dari 2 tahun lalu

it still doesn't work! After 45 minutes received this error: Error encountered while creating a cluster. Please try again or select an existing cluster. Failure Reason: Error while creating the new cluster - Error: Cluster not in running state

Paul M. · Diulas lebih dari 2 tahun lalu

Eduardo J. · Diulas lebih dari 2 tahun lalu

I coul not create any pipeline it gave me an error. I spent 2 hours waiting for the cluster to be created and no success

Niloofar M. · Diulas lebih dari 2 tahun lalu

Marco M. · Diulas lebih dari 2 tahun lalu

Aleksei K. · Diulas lebih dari 2 tahun lalu

Doesn't work the deploy of kubeflow pipelines. Creating "custer-1" in zone "us-central1-a" doesm't finish

Ivan R. · Diulas lebih dari 2 tahun lalu

Ivan R. · Diulas lebih dari 2 tahun lalu

Anas K. · Diulas lebih dari 2 tahun lalu

Kami tidak dapat memastikan bahwa ulasan yang dipublikasikan berasal dari konsumen yang telah membeli atau menggunakan produk terkait. Ulasan tidak diverifikasi oleh Google.