
Case No.PC-10086A customer who viewed the site asked us about a workstation for DeepLabCut ver2.3. We are looking for an optimal configuration with a budget of around 100 million yen.
Regarding the level of processing power, it is necessary to have a configuration that can run DeepLabCut at a speed equal to or faster than the workstation equipped with 6000 x NVIDIA RTX A1 that the customer is currently using, and we also request that DeepLabCut ver2.3 be installed in advance. It has received.
Other specific specifications are as follows.
CPU | Ability to run DeepLabCut without problems |
memory | 128GB |
GPU | Something that can perform DeepLabCut analysis at the same speed as RTX A6000 (Geforce RTX4090, etc.) |
storage | 1TB M.2 SSD x2 |
Based on the information received from the customer, we proposed the following configuration.
CPU | Core i9-14900K (3.20GHz 8 cores + 2.40GHz 16 cores) |
memory | 128GB |
Storage 1 | 1TB SSD M.2 NVMe Gen4 |
Storage 2 | 1TB SSD M.2 NVMe Gen4 |
Video | NVIDIA Geforce RTX4090 24GB |
network | on board (2.5GBase-T x1) Wi-Fi x1 |
Housing + power supply | Middle tower type housing + 1500W |
OS | Microsoft Windows 11 Professional 64bit |
Other | DeepLabCut installation |
This configuration is equipped with the latest 2024th generation Core i2 as of February 14. I chose the Core i9-24K with a total of 9 cores to suit my budget.
In addition, when actually executing DeepLabCut, most of the load is placed on the GPU, and it is thought that there are very few situations where the CPU is placed under a high load.
Therefore, if CPU performance is not important, you can change to a lower model such as Core i7.
GeForce RTX4090and RTX A6000, which one is best?
In accordance with the customer's wishes, we selected an NVIDIA GeForce RTX4090 24GB GPU.
Comparing NVIDIA GeForce RTX4090 24GB and NVIDIA RTX A6000 48GB, there are differences in chip generation and VRAM capacity. The chip generation used in the RTX4090 is newer, but the VRAM capacity is half that of the A6000.
DeepLabCut's GPU processing is based on TensorFlow, and its behavior while using GPU also follows the characteristics of TensorFlow. When executed with TensorFlow's default settings, DeepLabCut allocates as much video memory as possible and executes the process. Therefore, although there may be an effect due to the difference in VRAM capacity, the number of CUDA cores on the RTX4090 is over 18,000, which is nearly twice that of the A6000. In actual use, it is thought that the difference in the number of CUDA cores will have a greater effect than the difference in VRAM capacity, so we can expect an improvement in analysis speed compared to the RTX A2.
In addition, the GPU requirements posted in the DeepLabCut official repository recommend 8GB or more of VRAM. Considering that value, it can be said that the 4090GB of RTX24 has sufficient VRAM.
The configuration of this case study is based on the conditions given by the customer.
We will flexibly propose machines according to your conditions, so please feel free to contact us even if you are considering different conditions than what is listed.
■ Keywords・What is DeepLabCut? DeepLabCut is an open source deep learning tool for analyzing animal behavior. It can identify specific body parts of animals from videos and track them without markers, providing highly accurate movement analysis.
|
■ Click here for details and inquiries about this PC case * Please enter the name of the case or your desired conditions. |