Compute resources
Last updated
Was this helpful?
Last updated
Was this helpful?
Notebooks on Redivis provide a highly flexible computational environment. Notebooks can be used for anything from quick visualizations to training sophisticated ML models on a large corpus of data.
Understanding the compute resources available, and when to modify which parameters, can help you take full and efficient advantage of the high-performance computing resources on Redivis.
The default notebook configuration on Redivis is always free, and provides a performant environment for working with most datasets. The computational resources in the default notebook are comparable to a typical personal computer, though likely with substantially better network performance.
The default free notebook configuration offers:
2 vCPUs (Intel Ice Lake or Cascade Lake)
32GB RAM
100GB SSD:
IOPS: 170,000 read | 90,000 write
Throughput: 660MB/s read | 350MB/s write
16Gbps networking
No GPU (see below)
6 hr max duration
30min idle timeout (no code is being written or executed)
For scenarios where you need additional computational resources, you can choose a custom compute configuration for your notebook. This enables you to specify CPU, memory, GPU, and hard disk resources, while also giving you control over the notebook's max duration and idle timeout.
In order to customize the compute configuration for your notebook, click the Edit compute configuration button in the notebook start modal or toolbar.
These machine types are classified by four high-level compute platforms: General purpose, memory optimized, compute optimized, and GPU. Choose the platform, and machine type therein, that is mose appropriate for your workload.
All custom machines have an associated hourly cost (charged by the second). This cost is determined by the then-current price for that machine configuration on Google Cloud.
All notebooks on Redivis use either Python, R, Stata, or SAS. While Redivis notebooks are highly performant and scalable, the coding paradigms in these languages can introduce bottlenecks when working with very large tabular data. If you are running into issues with performance we suggest:
Adjust the compute resources in your notebook. This may help to resolve these bottlenecks depending on what is causing them!
< 1GB: probably doesn't matter, use what suits you!
1-10GB: probably fine for a notebook, though a transform might be faster.
10-100GB: maybe doable in a notebook, but you'll want to make sure to apply the right programming methodologies. Try to pre-cut your data if you can.
Redivis makes available nearly . These machines can scale from small servers all the way to massively powerful VMs with thousands of cores, terabytes of memory, and dozens of state-of-the-art GPUs.
In order to run a custom machine, you must first purchase , and have enough credits to run the notebook for at least 15 minutes. If you run low on credits and don't have configured, you will receive various alerts as your credits run low, and ultimately the notebook will shut down when you are out of credits.
Use to clean and reduce the size of your data before analyzing them further in a notebook. When possible, this will often be the most performant and cost-efficient approach.
Adjust your programming model to load data lazily or on-disk to exceeding memory limits. See our suggestions for working with larger tables in , , , and .
>100GB: You should probably cut the data first in a , unless if you really know what you're doing.