jaxjug July Meeting - Introduction to Spark

Event Details

jaxjug July Meeting  - Introduction to Spark

Time: July 8, 2015 from 6pm to 8pm
Location: Availity
Street: 10752 Deerwood Park Blvd S, Ste 110
City/Town: Jacksonville FL 32256
Website or Map: http://maps.google.com/maps?q…
Phone: Eyalwir@ yahoo.com
Event Type: meeting
Organized By: Eyal Wir
Latest Activity: Jul 7, 2015

Export to Outlook or iCal (.ics)

Event Description

Introduction to Spark

Presented by Carol McDonald, MapR Technologies



Apache Spark is a fast and general engine for large-scale data processing. In contrast to Hadoop's two-stage disk-based MapReduce paradigm, Spark's in-memory primitives provide performance up to 100 times faster for certain applications.



The Spark software stack includes a core data-proccessing engine, an interface for interactive querying, Sparkstreaming for streaming data analysis, and growing libraries for machine-learning and graph analysis. Spark is quickly establishing itself as a leading environment for doing fast, iterative in-memory and streaming analysis.



This talk will give an introduction the Spark stack, explain how Spark has lighting fast results, and how it complements Apache Hadoop.


Please RSVP!
http://www.meetup.com/Jacksonville-JAVA-User-Group-JaxJUG/events/223679551/

Comment Wall

Comment

RSVP for jaxjug July Meeting - Introduction to Spark to add comments!

Join Codetown

Might attend (1)

Happy 10th year, JCertif!

Notes

Welcome to Codetown!

Codetown is a social network. It's got blogs, forums, groups, personal pages and more! You might think of Codetown as a funky camper van with lots of compartments for your stuff and a great multimedia system, too! Best of all, Codetown has room for all of your friends.

When you create a profile for yourself you get a personal page automatically. That's where you can be creative and do your own thing. People who want to get to know you will click on your name or picture and…
Continue

Created by Michael Levin Dec 18, 2008 at 6:56pm. Last updated by Michael Levin May 4, 2018.

Looking for Jobs or Staff?

Check out the Codetown Jobs group.

 

Enjoy the site? Support Codetown with your donation.



InfoQ Reading List

Podcast: Transforming Life Sciences: AI, Vibe Coding, and Drug Development Acceleration

In this podcast, Shane Hastie, Lead Editor for Culture & Methods, spoke to Satish Kothapalli about the transformative impact of AI and vibe coding in life sciences software development, the acceleration of drug development timelines, and the evolving roles of developers in an AI-augmented environment.

By Satish Kothapalli

AWS Lambda Managed Instances: Serverless Flexibility Meets EC2 Cost Models

Unlock the power of AWS Lambda Managed Instances, seamlessly combining serverless functions with Amazon EC2 for optimal performance and cost efficiency. Designed for steady-state workloads, this solution automates instance management, reduces cold starts, and enables multi-concurrency.

By Steef-Jan Wiggers

Presentation: Hybrid Cloud-Native Networking in Enterprise - Some Assembly Required

Louis Ryan shares a compelling vision for modern cloud native hybrid networking. He critiques primitive network abstractions (the "Big IP" problem) and rigid security policies that rot and cause SPOFs. Discover how architects can elevate network functionality, bake in identity (mTLS/PKI), and leverage composability to achieve repeatable policy enforcement everywhere their applications run.

By Louis Ryan

Grab Adds Real-Time Data Quality Monitoring to Its Platform

Grab updated its internal platform to monitor Apache Kafka data quality in real time. The system uses FlinkSQL and an LLM to detect syntactic and semantic errors. It currently tracks 100+ topics, preventing invalid data from reaching downstream users. This proactive strategy aligns with industry trends to treat data streams as reliable products.

By Patrick Farry

NVIDIA Dynamo Addresses Multi-Node LLM Inference Challenges

Serving Large Language Models (LLMs) at scale is complex. Modern LLMs now exceed the memory and compute capacity of a single GPU or even a single multi-GPU node. As a result, inference workloads for 70B+, 120B+ parameter models, or pipelines with large context windows, require multi-node, distributed GPU deployments.

By Claudio Masolo

© 2025   Created by Michael Levin.   Powered by

Badges  |  Report an Issue  |  Terms of Service