Como executar jobs do Apache Spark no Cloud Dataproc avaliações
Carregando...
Nenhum resultado encontrado.

Aplique suas habilidades no console do Google Cloud

Como executar jobs do Apache Spark no Cloud Dataproc avaliações

45383 avaliações

Oren F. · Revisado há 9 months

Shankar P. · Revisado há 9 months

Manikandan M. · Revisado há 9 months

Prayoga S. · Revisado há 9 months

Srikanta P. · Revisado há 9 months

The project has been done

GIRIBABU M. · Revisado há 9 months

敬源 黃. · Revisado há 9 months

Latif I. · Revisado há 9 months

Jeferson Camilo S. · Revisado há 9 months

Jinou Y. · Revisado há 9 months

Unable to finish lab due to error Modified INPUT for LAB from pyspark.sql import SparkSession, SQLContext, Row gcs_bucket='[qwiklabs-gcp-01-31c23a3b2c2f]' spark = SparkSession.builder.appName("kdd").getOrCreate() sc = spark.sparkContext data_file = "gs://"+gcs_bucket+"//kddcup.data_10_percent.gz" raw_rdd = sc.textFile(data_file).cache() raw_rdd.take(5) OUTPUT: Setting default log level to "WARN". To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel). 24/11/18 23:53:00 INFO SparkEnv: Registering MapOutputTracker 24/11/18 23:53:00 INFO SparkEnv: Registering BlockManagerMaster 24/11/18 23:53:00 INFO SparkEnv: Registering BlockManagerMasterHeartbeat 24/11/18 23:53:00 INFO SparkEnv: Registering OutputCommitCoordinator --------------------------------------------------------------------------- IllegalArgumentException Traceback (most recent call last) /tmp/ipykernel_13667/2491634418.py in <cell line: 8>() 6 data_file = "gs://"+gcs_bucket+"//kddcup.data_10_percent.gz" 7 raw_rdd = sc.textFile(data_file).cache() ----> 8 raw_rdd.take(5) /usr/lib/spark/python/pyspark/rdd.py in take(self, num) 1848 """ 1849 items: List[T] = [] -> 1850 totalParts = self.getNumPartitions() 1851 partsScanned = 0 1852 /usr/lib/spark/python/pyspark/rdd.py in getNumPartitions(self) 597 2 598 """ --> 599 return self._jrdd.partitions().size() 600 601 def filter(self: "RDD[T]", f: Callable[[T], bool]) -> "RDD[T]": /opt/conda/miniconda3/lib/python3.10/site-packages/py4j/java_gateway.py in __call__(self, *args) 1319 1320 answer = self.gateway_client.send_command(command) -> 1321 return_value = get_return_value( 1322 answer, self.gateway_client, self.target_id, self.name) 1323 /usr/lib/spark/python/pyspark/sql/utils.py in deco(*a, **kw) 194 # Hide where the exception came from that shows a non-Pythonic 195 # JVM exception message. --> 196 raise converted from None 197 else: 198 raise IllegalArgumentException: java.net.URISyntaxException: Malformed IPv6 address at index 6: gs://[qwiklabs-gcp-01-31c23a3b2c2f]/kddcup.data_10_percent.gz

Richard S. · Revisado há 9 months

Syed Dameem K. · Revisado há 9 months

Çağrı K. · Revisado há 9 months

Anthony R. · Revisado há 9 months

Christopher H. · Revisado há 9 months

Azhar B. · Revisado há 9 months

Clément P. · Revisado há 9 months

César R. · Revisado há 9 months

Sanjay V. · Revisado há 9 months

Naethree P. · Revisado há 9 months

Mohan Babu N. · Revisado há 9 months

Ignacio G. · Revisado há 9 months

Juannean Y. · Revisado há 9 months

Creative P. · Revisado há 9 months

more examples using jobs from using service schedulers or schedule using cron from SSH

Jeinner Daniel B. · Revisado há 9 months

Não garantimos que as avaliações publicadas sejam de consumidores que compraram ou usaram os produtos. As avaliações não são verificadas pelo Google.