DatabeanStalk
  • DatabeanStalk documentation
  • DatabeanStalk Guides
    • Registration
      • Onboard to DatabeanStalk
      • Set up your DatabeanStalk account
    • DatabeanStalk Free Trial
      • Sign up
      • Login
    • Free Community Edition
      • Sign up
      • Login
    • Data Science Notebook
    • Data Analysis using Spark
  • Fundamentals
    • Projects
    • Members
    • Task Lists
    • Tasks
  • Use Cases
    • For Designers
      • Figma Integration
    • For Engineers
      • GitHub Integration
    • For Support
      • Intercom Integration
  • Extras
    • Keyboard Shortcuts
Powered by GitBook
On this page
  • Apache Spark Runtime Environment
  • Server Options
  1. DatabeanStalk Guides

Data Analysis using Spark

In this tutorial, you'll learn the basic steps to load and analyze data with Apache Spark in DatabeanStalk PySpark environment.

PreviousData Science NotebookNextProjects

Last updated 3 years ago

Apache Spark Runtime Environment

Click on JupyterHub from left menu or click on QuickStart Spark or PySpark, it will open up new secure window with your same user credentials

Server Options

Select server options for Run Managed Apache Spark environment and click on start.

Jupyter notebook start with multiple Spark runtime kernels like native python, PySpark, R or core Spark.

Initialize spark session with "sc" in notebook and Databeanstalk create Spark driver and two executors in kubernetes Databeanstalk plateform.