After performing exploratory analysis on your data, you'll begin the machine learning phase.
This phase involves writing code in notebooks. The supported code languages are Python and SQL.
Introducing Notebooks
A notebook is an interactive, web-based document that allows you to write and run code on your data, display the results, and combine with the code and results with explanatory text and images, all in one place.

The document is structured as a series of cells, where each cell contains either executable code or formatted text.
You can run these cells individually and in any order, allowing you to see the results of each step immediately without having to re-run an entire program.
Their step-by-step nature is perfect for a data science workflow: you can load data in one cell, clean it in the next, visualize it in another, and build a machine learning model further down. This makes it easy to experiment, debug, and see the immediate impact of your code.
Furthermore, because notebooks blend code with narrative text and visuals, they are an excellent tool for storytelling with data, creating reproducible research, and sharing your analysis with others in a format that is easy to read and understand.
Creating Notebooks in SAP Databricks
In SAP Databricks, you manage your notebooks in the Workspace section. You usually create a specific folder for notebooks to organize your workspace. You can also create subfolders for different projects.

When you create a notebook, the editor opens where you'll enter your code. This editor is like the one used for SQL queries, though the language differs. By default, the language is Python, but you can change it to SQL.

The data exploration and analysis phase could also be managed using a Python notebook instead of using SQL.