life is too short for a diary



Posts Tagged: pyspark

Debugging Spark Application Locally using remote container

One of the nifty feature in any code development is the ability to debug your application using break points. Submitting a Spark job while waiting for it to complete will waste a lot of time debugging. Spark jobs can be debugging with `break points` and `step over` and `step into` and `step out` commands...

Continue reading → spark pyspark docker vscode