Mapreducetrabajos

Filtro

Mis búsquedas recientes
Filtrar por:
Presupuesto
a
a
a
Tipo
Habilidades
Idiomas
    Estado del trabajo
    967 mapreduce trabajados encontrados, precios en USD

    consiste en desarrollar un esquema maestro-trabajador, como el esquema de procesamiento mas habitual en los entornos de computacion distribuida, parecido a modelos tan famosos como mapReduce

    $87 (Avg Bid)
    $87 Oferta promedio
    4 ofertas

    consiste en desarrollar un esquema maestro-trabajador, como el esquema de procesamiento mas habitual en los entornos de computacion distribuida, parecido a modelos tan famosos como mapReduce

    $51 (Avg Bid)
    $51 Oferta promedio
    1 ofertas
    Ingeniero de Datos Finalizado left

    Buscamos perfil profesional para desempeñarse como Ingeniero de Datos modalidad Freelance para proyecto en el rubro Minería. Excluyente experiencia de sobre 2 años como Desarrollador: Big Data Azure, Hadoop, MapReduce, Spark, Hive, Synapse Analytic, Python (o SCADA)

    $21 / hr (Avg Bid)
    $21 / hr Oferta promedio
    7 ofertas

    Me gustaría un curso de varias clases (cada clase un vídeo diferente) para aprender a utilizar Hadoop y MapReduce, al principio desde cero, e incrementando el nivel con cada clase, empezando con lo básico hasta los niveles más avanzados, para aprender y entender completamente Hadoop y MapReduce para Big Data. El mínimo total de duración sumando todas las clases que sea de 30 minutos. Importantísimo, idioma ESPAÑOL, sino no me sirve

    $8 - $31
    $8 - $31
    0 ofertas
    Big Data Architect Finalizado left

    Se trata de un proyecto generación de estrategia de Big Data para Business Analytics , las fases que queremos cubrir son las de revisión y auditoría de las fuentes de datos (estructurados - no estructurados) y el apoyo téc...sistemas operativos y redes. • Arquitectura de centro de procesos de datos y creación de Data Lakes (Cloudera,HortonWorks,MapR) • Familiarización con entornos de procesamientos modernos a escala masiva (Big Data) y/o en tiempo real: Hadoop/Mapreduce, HBase Scala/Spark, dataflow, Storm, Flume. • Conocimientos del entorno Salesforce •...

    $15 / hr (Avg Bid)
    $15 / hr Oferta promedio
    6 ofertas

    ...programación como Java, Scala o Python. Es imprescindible saber para qué y por qué se usan las tecnologías para poder modelar la mejor arquitectura posible para un determinado problema concreto de negocio. Buscamos: Titulados en Informática, Matemáticas, Estadística, etc Requisitos mínimos Experiencia con Git Lenguajes de programación: Java, Python, Scala, R... Experiencia en desarrollo de procesos MapReduce en Hadoop, Spark o Flink Manipulación de datos en diferentes DB Nosql como Cassandra, Mongo o HBase Requisitos deseados Experiencia en desarrollo de procesos Real Time con Storm, Spark o Flink Experiencia o conocimiento de las herramientas del ecosistema Hadoop Librerías y/o tecnologí...

    N/A
    N/A
    0 ofertas

    ...Profesional graduado en Informática, Economía, Actuario, Ciencias Exactas, oespecialistas en Data mining o Big Data.- Excelentes habilidades de comunicación, proactividad, capacidad de organización yplanificación.- Conocimientos de SQL.- Habilidad para crear modelos analíticos complejos y algoritmos.- Conocimientos de programación sobre herramientas y aplicaciones de Big Data (Hadoop/HDFS, Spark, MapReduce, Hive, R y Python)-Capacidad de abstracción y creatividad para resolver problemas complejos- Buena disposición tanto para el trabajo individual como para el trabajo en de trabajo: Saavedra, Capital :- Ambiente de trabajo desafiante y exigente pero a su vez alegre y divertido, para que hagas lo que más te gus...

    N/A
    N/A
    0 ofertas

    BluePatagon empresa líder en tecnologías de Business Intelligence & Business Analytics, está en búsqueda de Especialista en Big Data (Hadoop - Hortonworks), para importante cliente de CABA. Experiencia: mínima de 1 año en tecnologías Hadoop: Experiencia en desarrollo de aplicaciones (MapReduce + HDFS). Big Data: Familiaridad con el ecosistema (Hive, Pig, HBase, etc). y conceptos de escalabilidad, análisis en tiempo real, procesamiento de datos distribuidos. Linux: Uso Avanzado (manejo de servicios de SO, administración, shell scripting, seguridad) Programacion: POO (Java preferentemente, Python) Base de datos: RDBMS (Oracle, MySQL, PostgreSQL), NoSQL (HBase, Cassandra) Data Exchange y configuraci&oacu...

    N/A
    N/A
    0 ofertas

    El backend estaría compuesto por el lenguaje de programación Python, aunque también se programaría en Javascript haciendo uso de tecnologías como NodeJS.  La base de datos sería por un lado relacional PosgreSQL y por otro lado también utilizaríamo...relacional tipo MongoDB Para la plataforma en sí haríamos uso de tecnologías de frontend como AngularJS, HTML5, CSS3 etc. Básicamente es: lenguaje de programación PYTHON en el backend (en los servidores con los procesos de análisis de información) y JAVASCRIPT con el framework de AngularJS en la parte de cliente. Para el análisis de datos se utilizarán tecnologías usadas en Big Data como Hadoop, MapReduce, ...

    N/A
    N/A
    0 ofertas

    ...experiencia en NoSQL /HBASE, Cassandra o similares, Neo4j). Voluntad de aprender e implementar nuevas tecnologías BigData, según sea necesario. Iniciativa y capacidad de trabajar de manera independiente y en equipo. Experiencia con Sotrm en soluciones real time analytics. Experiencia en procesamiento paralelo (MPI, OpenMP) como ventaja competitiva para el puesto. Experto en comprensión en Hadoop HDFS y MapReduce. Pensamiento creativo (Out of the box) Capacidad en gestión de equipos <em>InnoQuant acaba de ser seleccionado como uno de los 10 startups tecnológicas más prometedoras en España.  Somos un equipo experimentado de profesionales de TI que trabajan en tiempo real plataforma de análisis de grandes datos del ...

    $244 (Avg Bid)
    $244 Oferta promedio
    5 ofertas

    given a dataset and using only MapReduce framework and python, find the following: • The difference between the maximum and the minimum for each day in the month • The daily minimum • the daily mean and variance • the correlation matrix that describes the monthly correlation among set of columns Using Mahout and python, do the following: • Implement the K-Means clustering algorithm • Find the optimum number (K) of clusters for the K-mean clustering • Plot the elbow graph for K-mean clustering • Compare the different clusters you obtained with different distance measures

    $165 (Avg Bid)
    $165 Oferta promedio
    8 ofertas

    Hello All, The objective of this subject is to learn how to design a distributed solution of a Big Data problem with help of MapReduce and Hadoop. In fact, MapReduce is a software framework for spreading a single computing job across multiple computers. It is assumed that these jobs take too long to run on a single computer, so you run them on multiple computers to shorten the time. Please stay auto bidders Thank You

    $100 (Avg Bid)
    $100 Oferta promedio
    4 ofertas
    MapReduce with Hadoop Finalizado left

    Hello All, The objective of this subject is to learn how to design a distributed solution of a Big Data problem with help of MapReduce and Hadoop. In fact, MapReduce is a software framework for spreading a single computing job across multiple computers. It is assumed that these jobs take too long to run on a single computer, so you run them on multiple computers to shorten the time. Please stay auto bidders Thank You

    $105 (Avg Bid)
    $105 Oferta promedio
    5 ofertas

    Hello All, The objective of this subject is to learn how to design a distributed solution of a Big Data problem with help of MapReduce and Hadoop. In fact, MapReduce is a software framework for spreading a single computing job across multiple computers. It is assumed that these jobs take too long to run on a single computer, so you run them on multiple computers to shorten the time. Please stay auto bidders Thank You

    $122 (Avg Bid)
    $122 Oferta promedio
    4 ofertas

    Hello All, The objective of this subject is to learn how to design a distributed solution of a Big Data problem with help of MapReduce and Hadoop. In fact, MapReduce is a software framework for spreading a single computing job across multiple computers. It is assumed that these jobs take too long to run on a single computer, so you run them on multiple computers to shorten the time. Please stay auto bidders Thank You

    $139 (Avg Bid)
    $139 Oferta promedio
    6 ofertas

    Hello All, The objective of this subject is to learn how to design a distributed solution of a Big Data problem with help of MapReduce and Hadoop. In fact, MapReduce is a software framework for spreading a single computing job across multiple computers. It is assumed that these jobs take too long to run on a single computer, so you run them on multiple computers to shorten the time. Please stay auto bidders Thank You

    $95 (Avg Bid)
    $95 Oferta promedio
    3 ofertas
    MapReduce with Hadoop Finalizado left

    The objective of this assignment is to learn how to design a distributed solution of a Big Data problem with help of MapReduce and Hadoop. In fact, MapReduce is a software framework for spreading a single computing job across multiple computers. It is assumed that these jobs take too long to run on a single computer, so you run them on multiple computers to shorten the time.

    $117 (Avg Bid)
    $117 Oferta promedio
    18 ofertas

    1. Implement the straggler solution using the approach below a) Develop a method to detect slow tasks (stragglers) in the Hadoop MapReduce framework using Progress Score (PS), Progress Rate (PR) and Remaining Time (RT) metrics b) Develop a method of selecting idle nodes to replicate detected slow tasks using the CPU time and Memory Status (MS) of the idle nodes. c) Develop a method for scheduling the slow tasks to appropriate idle nodes using CPU time and Memory Status of the idle nodes. 2. A good report on the implementation with graphics 3. A recorded execution process Use any certified data to test the efficiency of the methods

    $186 (Avg Bid)
    Urgente
    $186 Oferta promedio
    11 ofertas
    Big data project Finalizado left

    identify differences in implementations using Spark versus MapReduce, and understand LSH through implementing portions of the algorithm. Your task is to find hospitals with similar characteristics in the impact of COVID-19. Being able to quickly find similar hospitals can be useful for connecting hospitals experiencing difficulties and finding the characteristics of hospitals that have dealt better with the pandemic

    $189 (Avg Bid)
    $189 Oferta promedio
    17 ofertas
    mapreduce with python Finalizado left

    I have an input text file and a mapper and reducer file which outputs the total count of each word in the text file. I would like to have the mapper and reducer file output only the top 20 words (and their count) with the highest count. The files use and I wanna be able to run them in hadoop.

    $138 (Avg Bid)
    $138 Oferta promedio
    13 ofertas

    i want map reduce framework need to be implemented in scala

    $210 (Avg Bid)
    $210 Oferta promedio
    7 ofertas

    I will have couple of simple questions regarding: NLP, FSA, MapReduce, Regular expression, N-Gram. Please let me know if you have expertise in these topics.

    $158 (Avg Bid)
    $158 Oferta promedio
    35 ofertas

    1) Describe how to implement the following queries in MapReduce: SELECT , , , , FROM Employee as emp, Agent as a WHERE = AND = ; SELECT lo_quantity, COUNT(lo_extendedprice) FROM lineorder, dwdate WHERE lo_orderdate = d_datekey AND d_yearmonth = 'Feb1995' AND lo_discount = 6 GROUP BY lo_quantity; SELECT d_month, AVG(d_year) FROM dwdate GROUP BY d_month ORDER BY AVG(d_year) Consider a Hadoop job that processes an input data file of size equal to 179 disk blocks (179 different blocks, not considering HDFS replication factor). The mapper in this job requires 1 minute to read and fully process a single block of data. Reducer requires 1 second (not minute) to produce an answer for one key worth of values and there are a total of

    $200 (Avg Bid)
    $200 Oferta promedio
    1 ofertas

    I can successfully run the Mapreduce job on the server. But when I want to send this job as yarn remote client with java(via yarn Rest api), I get the following error. I want to submit this job successfully via Remote Client(Yarn Rest Api.)

    $12 (Avg Bid)
    $12 Oferta promedio
    3 ofertas

    Write a MapReduce program with python to implement BFS. , and shell script are needed according to the detailed instructions in the uploaded file.

    $136 (Avg Bid)
    $136 Oferta promedio
    24 ofertas

    ...to you how you pick necessary features and build the training that creates matching courses for job profiles. These are the suggested steps you should follow : Step 1: Setup a Hadoop cluster where the data sets should be stored on the set of Hadoop data nodes. Step 2: Implement a content based recommendation system using MapReduce, i.e. given a job description you should be able to suggest a set of applicable courses. Step 3: Execute the training step of your MapReduce program using the data set stored in the cluster. You can use a subset of the data depending on the system capacity of your Hadoop cluster. You have to use an appropriate subset of features in the data set for effective training. Step 4: Test your recommendation system using a set of requests that execute ...

    $20 (Avg Bid)
    $20 Oferta promedio
    3 ofertas

    The write-up should include the main problem that can be subdivided into 3 or 4 subproblems. If I'm satisfied, we discuss further on implementation.

    $235 (Avg Bid)
    $235 Oferta promedio
    4 ofertas

    Using mapreduce recommend the best courses for up-skilling based on a given job description. You can use the data set to train the system and pick some job descriptions not in the training set to test. It is left up to you how you pick necessary features and build the training that creates matching courses for job profiles. Project submission- 1. Code files with comments for your MapReduce implementation of training and query steps 2. . Document the design of your logic including training, query and feature engineering. data csv is too big will share separately

    $79 (Avg Bid)
    $79 Oferta promedio
    2 ofertas

    Hi Sri Varadan Designers, I noticed your profile and would like to offer you my project. We can discuss any details over chat. I have task to do in Mapreduce in hadoop

    $16 (Avg Bid)
    $16 Oferta promedio
    1 ofertas

    I want to run pouchdb-node on AWS Lambda. Source code: Detailed Requirements: - Deploy pouchdb-node to AWS Lambda. - Use EFS in storage layer. - Ok to limit concurrency to 1 to avoid race conditions. - Expose via Lambda HTTPS Endpoints (no API Gateway) - The basic PUT / GET functions, replication, and MapReduce must all work Project Deliverables: - Deployment script which packages pouchdb-node and deploys it to AWS using SAM or CloudFormation. Development Process: - I will not give access to my AWS Accounts. - You develop on your own environment and give me completed solution.

    $187 (Avg Bid)
    $187 Oferta promedio
    8 ofertas
    Hadoop Assignment Finalizado left

    ...to you how you pick necessary features and build the training that creates matching courses for job profiles. These are the suggested steps you should follow : Step 1: Setup a Hadoop cluster where the data sets should be stored on the set of Hadoop data nodes. Step 2: Implement a content based recommendation system using MapReduce, i.e. given a job description you should be able to suggest a set of applicable courses. Step 3: Execute the training step of your MapReduce program using the data set stored in the cluster. You can use a subset of the data depending on the system capacity of your Hadoop cluster. You have to use an appropriate subset of features in the data set for effective training. Step 4: Test your recommendation system using a set of requests that execute ...

    $130 (Avg Bid)
    $130 Oferta promedio
    5 ofertas

    ...metrics to show which is a better method. OR ii) Improvement on the methodology used in (a) that will produce a better result. 2. Find a suitable paper on replication of data in hadoop mapreduce framework. a) Implement the methodology used in the paper b) i) Write a program to split identified intermediate results from (1 b(i)) appropriately into 64Mb/128Mb and compare with 2(a) using same metrics to show which is a better method. OR ii) Improvement on the methodology used in 2(a) that will produce a better result 3. Find a suitable paper on allocation strategy of data/tasks to nodes in Hadoop Mapreduce framework. a) Implement the methodology used in the paper b) i) Write a program to reallocate the splits from (2 (b(i)) above to nodes by considering the capability ...

    $158 (Avg Bid)
    $158 Oferta promedio
    5 ofertas

    ... SQL Concepts, Data Modelling Techniques & Data Engineering Concepts is a must Hands on experience in ETL process, Performance optimization techniques is a must. Candidate should have taken part in Architecture design and discussion. Minimum of 2 years of experience in working with batch processing/ real-time systems using various technologies like Databricks, HDFS, Redshift, Hadoop, Elastic MapReduce on AWS, Apache Spark, Hive/Impala and HDFS, Pig, Kafka, Kinesis, Elasticsearch and NoSQL databases Minimum of 2 years of experience working in Datawarehouse or Data Lake Projects in a role beyond just Data consumption. Minimum of 2 years of extensive working knowledge in AWS building scalable solutions. Equivalent level of experience in Azure or Google Cloud is also acceptable M...

    $1864 (Avg Bid)
    $1864 Oferta promedio
    16 ofertas
    Project for Mohd T. Finalizado left

    Hi Mohd. I hope you are well, I have some Big Data exercises (hive, pig, sed and mapreduce) I would like to know if you can help me

    $80 (Avg Bid)
    $80 Oferta promedio
    1 ofertas

    1) Develop an aggregate of these reviews using your knowledge of Hadoop and MapReduce in Microsoft HDInsight. a) Follow the same approach as the Big Data Analytics Workshop (using the wordcount method in HDInsight) to determine the contributory words for each level of rating. b) Present the workflow of using HDInsight (you may use screen captures) along with a summary of findings and any insights for each level of rating. MapReduce documentation for HDInsight is available here 2) Azure data bricks for some insights Provide the following: a) A screen capture of the completed model diagram and any decision you made in training the model. For example, rationale for some of the components used, how many records have been used for training and how many for testing. b) A set of ...

    $147 (Avg Bid)
    $147 Oferta promedio
    7 ofertas

    I am looking for a java developer who is -familiar with hadoop architecture and mapreduce scheduling -familiar with modifying the open source packages

    $250 (Avg Bid)
    $250 Oferta promedio
    5 ofertas
    Big data project Finalizado left

    ...7910/DVN/HG7NV7 4. Design, implement and run an Oozie workflow to find out a. the 3 airlines with the highest and lowest probability, respectively, of being on schedule; b. the 3 airports with the longest and shortest average taxi time per flight (both in and out), respectively; and c. the most common reason for flight cancellations. • Requirements: 1. Your workflow must contain at least three MapReduce jobs that run in fully distributed mode. 2. Run your workflow to analyze the entire data set (total 22 years from 1987 to 2008) at one time on two VMs first and then gradually increase the system scale to the maximum allowed number of VMs for at least 5 increment steps, and measure each corresponding workflow execution time. 3. Run your workflow to analyze the data in a prog...

    $202 (Avg Bid)
    $202 Oferta promedio
    7 ofertas
    data scientist neaded Finalizado left

    ...7910/DVN/HG7NV7 4. Design, implement and run an Oozie workflow to find out a. the 3 airlines with the highest and lowest probability, respectively, of being on schedule; b. the 3 airports with the longest and shortest average taxi time per flight (both in and out), respectively; and c. the most common reason for flight cancellations. • Requirements: 1. Your workflow must contain at least three MapReduce jobs that run in fully distributed mode. 2. Run your workflow to analyze the entire data set (total 22 years from 1987 to 2008) at one time on two VMs first and then gradually increase the system scale to the maximum allowed number of VMs for at least 5 increment steps, and measure each corresponding workflow execution time. 3. Run your workflow to analyze the data in a prog...

    $21 (Avg Bid)
    $21 Oferta promedio
    4 ofertas

    ...7910/DVN/HG7NV7 4. Design, implement and run an Oozie workflow to find out a. the 3 airlines with the highest and lowest probability, respectively, of being on schedule; b. the 3 airports with the longest and shortest average taxi time per flight (both in and out), respectively; and c. the most common reason for flight cancellations. • Requirements: 1. Your workflow must contain at least three MapReduce jobs that run in fully distributed mode. 2. Run your workflow to analyze the entire data set (total 22 years from 1987 to 2008) at one time on two VMs first and then gradually increase the system scale to the maximum allowed number of VMs for at least 5 increment steps, and measure each corresponding workflow execution time. 3. Run your workflow to analyze the data in a prog...

    $11 (Avg Bid)
    $11 Oferta promedio
    4 ofertas
    data project Finalizado left

    ...7910/DVN/HG7NV7 4. Design, implement and run an Oozie workflow to find out a. the 3 airlines with the highest and lowest probability, respectively, of being on schedule; b. the 3 airports with the longest and shortest average taxi time per flight (both in and out), respectively; and c. the most common reason for flight cancellations. • Requirements: 1. Your workflow must contain at least three MapReduce jobs that run in fully distributed mode. 2. Run your workflow to analyze the entire data set (total 22 years from 1987 to 2008) at one time on two VMs first and then gradually increase the system scale to the maximum allowed number of VMs for at least 5 increment steps, and measure each corresponding workflow execution time. 3. Run your workflow to analyze the data in a prog...

    $144 (Avg Bid)
    $144 Oferta promedio
    6 ofertas
    data progect Finalizado left

    ...7910/DVN/HG7NV7 4. Design, implement and run an Oozie workflow to find out a. the 3 airlines with the highest and lowest probability, respectively, of being on schedule; b. the 3 airports with the longest and shortest average taxi time per flight (both in and out), respectively; and c. the most common reason for flight cancellations. • Requirements: 1. Your workflow must contain at least three MapReduce jobs that run in fully distributed mode. 2. Run your workflow to analyze the entire data set (total 22 years from 1987 to 2008) at one time on two VMs first and then gradually increase the system scale to the maximum allowed number of VMs for at least 5 increment steps, and measure each corresponding workflow execution time. 3. Run your workflow to analyze the data in a prog...

    $10 (Avg Bid)
    $10 Oferta promedio
    3 ofertas

    Familiarity with Hadoop ecosystem and its components: obviously, a must! Ability to write reliable, manageable, and high-performance code Expertise knowledge of Hadoop HDFS, Hive, Pig, Flume and Sqoop. Working experience in HQL Experience of writing Pig Latin and MapReduce jobs Good knowledge of the concepts of Hadoop. Analytical and problem-solving skills; the implementation of these skills in Big Data domain Understanding of data loading tools such as Flume, Sqoop etc Good knowledge of database principles, practices, structures, and theories

    $613 (Avg Bid)
    $613 Oferta promedio
    2 ofertas

    Using ansible, harvest twitter data with geo coordinates using twitter API and put into a couchDB. The CouchDB setup may be a single node or based on a cluster setup. The cloud based solution should use 4 VMs with 8 virtual CPUs and 500Gb of volume storage. The data is then combined with other useful geographic data to produce some visualization summary results using MapReduce.

    $101 (Avg Bid)
    $101 Oferta promedio
    8 ofertas

    Write a MapReduce program to analyze the income data extracted from the 1990 U.S. Census data and determine whether most Americans make more than $50,000 or $50,000 or less a year in 1990. Provide the number of people who made more than $50,000 and the number of people who made $50,000 or less. Download data from http://archive.ics.uci.edu/ml/datasets/Census+Income

    $162 (Avg Bid)
    Urgente
    $162 Oferta promedio
    7 ofertas

    1 Explain the concept of Big Data and its importance in a modern economy 2 Explain the core architecture and algorithms underpinning big data processing 3 Analyse and visualize large data sets using a range of statistical and big data technologies 4 Critically evaluate, select and employ appropriate tools and technologies for the development of big data applications

    $18 - $153
    Sellado Acuerdo de Confidencialidad
    $18 - $153
    2 ofertas

    Big Data task with the use of python and hadoop using mapreduce techniques

    $14 (Avg Bid)
    $14 Oferta promedio
    6 ofertas
    big data management Finalizado left

    Hadoop, Implementation of MapReduce application

    $15 (Avg Bid)
    $15 Oferta promedio
    7 ofertas

    Parsing, Cleaning, and Profiling of the attached file by removing hashtags, emoticons, or any redundant data which is not useful for analysis. And MapReduce output will be on HDFS like the image attached named "Output" but should be clean. Tasks: Dataset: Programming: MapReduce with Java Data profiling: Write MapReduce java code to characterize (profile) the data in each column. Data cleaning: Cleaning and Profiling the tweets by removing hashtags, emoticons, or any redundant data which is not useful for analysis. Write MapReduce java code to ETL (extract, transform, load) data source. Drop some unimportant columns, Normalize data in a column, and Detect badly formatted rows.

    $20 (Avg Bid)
    $20 Oferta promedio
    1 ofertas

    ...con l’architettura utilizzata in tutta l’azienda. Competenze richieste - Laurea in Informatica, Information Technology o equivalente esperienza tecnica. - Almeno 3 anni di esperienza professionale. - Profonda conoscenza ed esperienza in statistica. - Previa esperienza in programmazione, preferibilmente in Python, Kafka o Java e volontà di apprende nuovi linguaggi. - Competenze su Hadoop v2, MapReduce, HDFS. - Buona conoscenza dei Big Data querying tools. - Esperienza con Spark. -Esperienza nel processare grandi quantità di dati, sia strutturati che non, inclusa l’integrazione di dati che provengono da fonti diverse. - Esperienza con NoSQL databases, come Cassandra o MongoDB. - Esperienza con vari sistemi di messagistica, come Kafka o RabbitMQ Du...

    $22 / hr (Avg Bid)
    $22 / hr Oferta promedio
    6 ofertas
    Hadoop - Mapreduce Finalizado left

    I need some help with a small task completing some beginning steps in Hadoop with python. Come to the chat and I can explain more. It will not take long, the only thing you need is virtualbox and some som python & Hadoop knowledge.

    $21 (Avg Bid)
    $21 Oferta promedio
    4 ofertas

    Principales artículos de la comunidad mapreduce