Favoring standard libraries over external dependencies, especially in specific contexts like Databricks, is a best practice in software development.
There are several reasons why this approach is encouraged:
- Standard libraries are typically well-vetted, thoroughly tested, and maintained by the official maintainers of the programming language or platform. This ensures a higher level of stability and reliability.
- External dependencies, especially lesser-known or unmaintained ones, can introduce bugs, security vulnerabilities, or compatibility issues that can be challenging to resolve. Adding external dependencies increases the complexity of your codebase.
- Each dependency may have its own set of dependencies, potentially leading to a complex web of dependencies that can be difficult to manage. This complexity can lead to maintenance challenges, increased risk, and longer build times.
- External dependencies can pose security risks. If a library or package has known security vulnerabilities and is widely used, it becomes an attractive target for attackers. Minimizing external dependencies reduces the potential attack surface and makes it easier to keep your code secure.
- Relying on standard libraries enhances code portability. It ensures your code can run on different platforms and environments without being tightly coupled to specific external dependencies. This is particularly important in settings like Databricks, where you may need to run your code on different clusters or setups.
- External dependencies may have their versioning schemes and compatibility issues. When using standard libraries, you have more control over versioning and can avoid conflicts between different dependencies in your project.
- Fewer external dependencies mean faster build and deployment times. Downloading, installing, and managing external packages can slow down these processes, especially in large-scale projects or distributed computing environments like Databricks.
- External dependencies can be abandoned or go unmaintained over time. This can lead to situations where your project relies on outdated or unsupported code. When you depend on standard libraries, you have confidence that the core functionality you rely on will continue to be maintained and improved.
While minimizing external dependencies is essential, exceptions can be made case-by-case. There are situations where external dependencies are justified, such as when a well-established and actively maintained library provides significant benefits, like time savings, performance improvements, or specialized functionality unavailable in standard libraries.
The Remorph project requires any commit to be signed-off using GPG signing. Before you submit any commit, please make sure you are properly setup, as follows.
If you don't already have one, create a GPG key:
- on MacOS, install the GPG Suite from https://gpgtools.org/
- from the Applications folder, launch the GPG Keychain app
- create a new GPG key, using your Databricks email
- Right-click on the created key and select Export, to save the key
- Check the key using TextEdit, it should start with -----BEGIN PGP PUBLIC KEY BLOCK-----
Register your PGP key in GitHub:
- In GitHub, select Settings from your picture at the top-right
- select SSH and PGP key
- click on New key, and paste the text content of the exported key
- select Emails
- if your databricks email is not registered, register it
- complete the verification before the next steps
- select SSH and PGP key
Tell local git to signoff your commits using your PGP key - see full instructions here https://docs.github.com/en/authentication/managing-commit-signature-verification/telling-git-about-your-signing-key - in short, you need to run the following commands from a terminal: - git config --global --unset gpg.format - gpg --list-secret-keys --keyid-format=long - git config --global user.signingkey <KEY.ID.FROM.ABOVE> - git config --global commit.gpgsign true
Once all this is done, you can verify it's correct as follows:
- create a branch and use it
- create a file with some content
- git add
- git commit -m "test PGP"
- git verify-commit
The last command should display something like the following:
gpg: Signature made Tue Nov 26 11:34:23 2024 CET gpg: using RSA key FD4D754BB2B1D4F09F2BF658F4B0C73DFC65A17B gpg: Good signature from "GitHub <[email protected]>" [ultimate]
When you introduce a change in the code, specifically a deeply technical one, please ensure that the change provides same or improved set of capabilities. PRs that remove existing functionality shall be properly discussed and justified.
When writing code, divide it into two main parts: Components for API Interaction and Components for Business Logic. API Interaction should only deal with talking to external systems through APIs. They are usually integration-tested, and mocks are simpler. Business Logic handles the actual logic of your application, like calculations, data processing, and decision-making.
Keep API components simple. In the components responsible for API interactions, try to keep things as straightforward as possible. Refrain from overloading them with complex logic; instead, focus on making API calls and handling the data from those calls.
Inject Business Logic. If you need to use business logic in your API-calling components, don't build it directly there. Instead, inject (or pass in) the business logic components into your API components. This way, you can keep your API components clean and flexible, while the business logic remains separate and reusable.
Test your Business Logic. It's essential to test your business logic to ensure it works correctly and thoroughly. When writing unit tests, avoid making actual API calls - unit tests are executed for every pull request, and take seconds to complete. For calling any external services, including Databricks Connect, Databricks Platform, or even Apache Spark, unit tests have to use "mocks" or fake versions of the APIs to simulate their behavior. This makes testing your code more manageable and catching any issues without relying on external systems. Focus on testing the edge cases of the logic, especially the scenarios where things may fail. See this example as a reference of an extensive unit test coverage suite and the clear difference between unit tests and integration tests.
In order to use this, you have to install remorph
on any workspace via databricks labs install .
,
so that .databricks-login.json
file gets created with the following contents:
{
"workspace_profile": "labs-azure-tool",
"cluster_id": "0708-200540-wcwi4i9e"
}
then run make dev-cli
to collect classpath information. And then invoke commands,
like databricks labs remorph debug-script --name file
. Add --debug
flag to recompile project each run.
Example output is:
databricks labs remorph debug-script --name foo
21:57:42 INFO [databricks.sdk] Using Azure CLI authentication with AAD tokens
21:57:42 WARN [databricks.sdk] azure_workspace_resource_id field not provided. It is recommended to specify this field in the Databricks configuration to avoid authentication errors.
Debugging script...
Map(log_level -> disabled, name -> foo)
This section provides a step-by-step guide to set up and start working on the project. These steps will help you set up your project environment and dependencies for efficient development.
To begin, install prerequisites:
wget
is required by the maven installer
brew install wget
maven
is the dependency manager for JVM based languages
brew install maven
jdk11
is the jdk used by remorph
download it from OpenJDK11 and install it
python
is the dependency manager for JVM based languages
brew install maven
hatch
is a Python project manager
pip install hatch
Then run project-specific install scripts
make dev
creates the default environment and installs development dependencies, assuming you've already cloned the github repo.
make dev
Verify installation with
make test
To ensure your integrated development environment (IDE) uses the newly created virtual environment, you can retrieve the Python path with this command:
hatch run python -c "import sys; print(sys.executable)"
As of writing, we only support IntelliJ IDEA CE 2024.1. Development using more recent versions doesn't work (yet!). Download and install IntelliJ IDEA
Configure your IDE to:
- use OpenJDK11 as the SDK for the project
- install the IntelliJ Scala plugin version 2024.1.25. Do not use more recent versions, they don't work!!!
- use this Python venv path so that you work within the virtual environment when developing the project:
Before every commit, apply the consistent formatting of the code, as we want our codebase look consistent:
make fmt
Before every commit, run automated bug detector (make lint
) and unit tests (make test
) to ensure that automated
pull request checks do pass, before your code is reviewed by others:
make lint test
If you will be working with the ANTLR grammars, then you should install the ANTLR plugin for your IDE. There is a plugin for VS Code, but it does not have as many checks as the one for IntelliJ IDEA.
While the ANTLR tool run at build time, will warn (and the build will stop on warnings) about things like tokens that are used in the parser grammar but not defined in the lexer grammar, the IntelliJ IDEA plugin provides a few extra tools such as identifying unused rules, and providing a visual representation of trees etc.
Please read the documentation for the plugin so that you can make the most of it and have it generate the lexer and parser in a temp directory to tell you about things like undefined tokens etc.
If you intended to make changes to the ANTLR defined syntax, please read teh README.md under ./core before doing so. Changes to ANTLR grammars can have a big knock on effects on the rest of the codebase, and must be carefully reviewed and tested - all the way from parse to code generation. Such changes are generally not suited to beginners.
Here are the example steps to submit your first contribution:
- Make a Fork from remorph repo (if you really want to contribute)
git clone
git checkout main
(orgcm
if you're using ohmyzsh).git pull
(orgl
if you're using ohmyzsh).git checkout -b FEATURENAME
(orgcb FEATURENAME
if you're using ohmyzsh).- .. do the work
make fmt
make lint
- .. fix if any
make test
- .. fix if any
git commit -a
. Make sure to enter meaningful commit message title.git push origin FEATURENAME
- Go to GitHub UI and create PR. Alternatively,
gh pr create
(if you have GitHub CLI installed). Use a meaningful pull request title because it'll appear in the release notes. UseResolves #NUMBER
in pull request description to automatically link it to an existing issue. - announce PR for the review
If you encounter any package dependency errors after git pull
, run make clean