What happens if I start too many background jobs?Where do background jobs go?Is it possible to customise the...
Is this homebrew race based on Draco Volans balanced?
Is it cheaper to drop cargo than to land it?
Is thermodynamics only applicable to systems in equilibrium?
What happened to Ghost?
Problems with numbers (result of calculations) alignment using siunitx package inside tabular environment
Can I use 1000v rectifier diodes instead of 600v rectifier diodes?
Sower of Discord, Gideon's Sacrifice and Stuffy Doll
Airbnb - host wants to reduce rooms, can we get refund?
Visualizing a complicated Region
Who died in the Game of Thrones episode, "The Long Night"?
Why is the SNP putting so much emphasis on currency plans?
Declining welcome lunch invitation at new job due to Ramadan
How did Arya manage to disguise herself?
Can a cyclic Amine form an Amide?
If Earth is tilted, why is Polaris always above the same spot?
Unidentified items in bicycle tube repair kit
How did Captain America use this power?
If Melisandre foresaw another character closing blue eyes, why did she follow Stannis?
Stark VS Thanos
Pressure to defend the relevance of one's area of mathematics
Why was the pattern string not followed in this code?
How to back up a running Linode server?
Copy line and insert it in a new position with sed or awk
Was the ancestor of SCSI, the SASI protocol, nothing more than a draft?
What happens if I start too many background jobs?
Where do background jobs go?Is it possible to customise the prompt to show the if there are any background jobs?What happens to suspended jobs in unix?Modify background jobs and/or notify on job completion post-launchLive monitoring of background jobsAre background jobs and foreground jobs always in running state?What happens to background jobs after exiting the shell?How do I wait for background jobs in the background?how to wait for many background jobs in bashHow can we kill all the background jobs (running or not) in the current shell?
.everyoneloves__top-leaderboard:empty,.everyoneloves__mid-leaderboard:empty,.everyoneloves__bot-mid-leaderboard:empty{ margin-bottom:0;
}
I need to do some work on 700 network devices using an expect script. I can get it done sequentially, but so far the runtime is around 24 hours. This is mostly due to the time it takes to establish a connection and the delay in the output from these devices (old ones). I'm able to establish two connections and have them run in parallel just fine, but how far can I push that?
I don't imagine I could do all 700 of them at once, surely there's some limit to the no. of telnet connections my VM can manage.
If I did try to start 700 of them in some sort of loop like this:
for node in `ls ~/sagLogs/`; do
foo &
done
With
CPU 12 CPUs x Intel(R) Xeon(R) CPU E5649 @ 2.53GHz
Memory 47.94 GB
My question is:
- Could all 700 instances possibly run concurrently?
- How far could I get until my server reaches its limit?
- When that limit is reached, will it just wait to begin the next iteration off
foo
or will the box crash?
I'm running in a corporate production environment unfortunately, so I can't exactly just try and see what happens.
bash background-process expect telnet jobs
add a comment |
I need to do some work on 700 network devices using an expect script. I can get it done sequentially, but so far the runtime is around 24 hours. This is mostly due to the time it takes to establish a connection and the delay in the output from these devices (old ones). I'm able to establish two connections and have them run in parallel just fine, but how far can I push that?
I don't imagine I could do all 700 of them at once, surely there's some limit to the no. of telnet connections my VM can manage.
If I did try to start 700 of them in some sort of loop like this:
for node in `ls ~/sagLogs/`; do
foo &
done
With
CPU 12 CPUs x Intel(R) Xeon(R) CPU E5649 @ 2.53GHz
Memory 47.94 GB
My question is:
- Could all 700 instances possibly run concurrently?
- How far could I get until my server reaches its limit?
- When that limit is reached, will it just wait to begin the next iteration off
foo
or will the box crash?
I'm running in a corporate production environment unfortunately, so I can't exactly just try and see what happens.
bash background-process expect telnet jobs
I’m guessing each job uses very little CPU and RAM, is that right?
– Stephen Kitt
3 hours ago
Honestly I have a hard time telling. Htop isn't very helpful - when I'm running one instance the CPU reads:CPU: 86.9% sys: 13.1% low: 0.0%
and RAM readsMem:3.86G used:178M buffers:2.28G cache:608M
Any guess?
– KuboMD
3 hours ago
add a comment |
I need to do some work on 700 network devices using an expect script. I can get it done sequentially, but so far the runtime is around 24 hours. This is mostly due to the time it takes to establish a connection and the delay in the output from these devices (old ones). I'm able to establish two connections and have them run in parallel just fine, but how far can I push that?
I don't imagine I could do all 700 of them at once, surely there's some limit to the no. of telnet connections my VM can manage.
If I did try to start 700 of them in some sort of loop like this:
for node in `ls ~/sagLogs/`; do
foo &
done
With
CPU 12 CPUs x Intel(R) Xeon(R) CPU E5649 @ 2.53GHz
Memory 47.94 GB
My question is:
- Could all 700 instances possibly run concurrently?
- How far could I get until my server reaches its limit?
- When that limit is reached, will it just wait to begin the next iteration off
foo
or will the box crash?
I'm running in a corporate production environment unfortunately, so I can't exactly just try and see what happens.
bash background-process expect telnet jobs
I need to do some work on 700 network devices using an expect script. I can get it done sequentially, but so far the runtime is around 24 hours. This is mostly due to the time it takes to establish a connection and the delay in the output from these devices (old ones). I'm able to establish two connections and have them run in parallel just fine, but how far can I push that?
I don't imagine I could do all 700 of them at once, surely there's some limit to the no. of telnet connections my VM can manage.
If I did try to start 700 of them in some sort of loop like this:
for node in `ls ~/sagLogs/`; do
foo &
done
With
CPU 12 CPUs x Intel(R) Xeon(R) CPU E5649 @ 2.53GHz
Memory 47.94 GB
My question is:
- Could all 700 instances possibly run concurrently?
- How far could I get until my server reaches its limit?
- When that limit is reached, will it just wait to begin the next iteration off
foo
or will the box crash?
I'm running in a corporate production environment unfortunately, so I can't exactly just try and see what happens.
bash background-process expect telnet jobs
bash background-process expect telnet jobs
asked 3 hours ago
KuboMDKuboMD
1888
1888
I’m guessing each job uses very little CPU and RAM, is that right?
– Stephen Kitt
3 hours ago
Honestly I have a hard time telling. Htop isn't very helpful - when I'm running one instance the CPU reads:CPU: 86.9% sys: 13.1% low: 0.0%
and RAM readsMem:3.86G used:178M buffers:2.28G cache:608M
Any guess?
– KuboMD
3 hours ago
add a comment |
I’m guessing each job uses very little CPU and RAM, is that right?
– Stephen Kitt
3 hours ago
Honestly I have a hard time telling. Htop isn't very helpful - when I'm running one instance the CPU reads:CPU: 86.9% sys: 13.1% low: 0.0%
and RAM readsMem:3.86G used:178M buffers:2.28G cache:608M
Any guess?
– KuboMD
3 hours ago
I’m guessing each job uses very little CPU and RAM, is that right?
– Stephen Kitt
3 hours ago
I’m guessing each job uses very little CPU and RAM, is that right?
– Stephen Kitt
3 hours ago
Honestly I have a hard time telling. Htop isn't very helpful - when I'm running one instance the CPU reads:
CPU: 86.9% sys: 13.1% low: 0.0%
and RAM reads Mem:3.86G used:178M buffers:2.28G cache:608M
Any guess?– KuboMD
3 hours ago
Honestly I have a hard time telling. Htop isn't very helpful - when I'm running one instance the CPU reads:
CPU: 86.9% sys: 13.1% low: 0.0%
and RAM reads Mem:3.86G used:178M buffers:2.28G cache:608M
Any guess?– KuboMD
3 hours ago
add a comment |
2 Answers
2
active
oldest
votes
Could all 700 instances possibly run concurrently?
That depends on what you mean by concurrently. If we're being picky, then no, they can't unless you have 700 threads of execution on your system you can utilize (so probably not). Realistically though, yes, they probably can, provided you have enough RAM and/or swap space on the system. UNIX and it's various children are remarkably good at managing huge levels of concurrency, that's part of why they're so popular for large-scale HPC usage.
How far could I get until my server reaches its limit?
This is impossible to answer concretely without a whole lot more info. Pretty much, you need to have enough memory to meet:
- The entire run-time memory requirements of one job, times 700.
- The memory requirements of bash to manage that many jobs (bash is not horrible about this, but the job control isn't exactly memory efficient).
- Any other memory requirements on the system.
Assuming you meet that (again, with only 50GB of RAM, you still ahve to deal with other issues:
- How much CPU time is going to be wasted by bash on job control? Probably not much, but with hundreds of jobs, it could be significant.
- How much network bandwidth is this going to need? Just opening all those connections may swamp your network for a couple of minutes depending on your bandwidth and latency.
- Many other things I probably haven't thought of.
When that limit is reached, will it just wait to begin the next iteration off foo or will the box crash?
It depends on what limit is hit. If it's memory, something will die on the system (more specifically, get killed by the kernel in an attempt to free up memory) or the system itself may crash (it's not unusual to configure systems to intentionally crash when running out of memory). If it's CPU time, it will just keep going without issue, it'll just be impossible to do much else on the system. If it's the network though, you might crash other systems or services.
What you really need here is not to run all the jobs at the same time. Instead, split them into batches, and run all the jobs within a batch at the same time, let them finish, then start the next batch. GUN Parallel can be used for this, but it's less than ideal at that scale in a production environment (if you go with it, don't get too aggressive, like I said, you might swamp the network and affect systems you otherwise would not be touching). I would really recommend looking into a proper network orchestration tool like Ansible (https://www.ansible.com/), as that will not only solve your concurrency issues (Ansible does batching like I mentioned above automatically), but also give you a lot of other useful features to work with (like idempotent execution of tasks, nice status reports, and native integration with a very large number of other tools).
add a comment |
It's hard to say specifically how many instances could be run as background jobs in the manner you describe. But a normal server can certainly maintain 700 concurrent connections as long as you do it correctly. Webservers do this all the time.
May I suggest that you use GNU parallel (https://www.gnu.org/software/parallel/) or something similar to accomplish this? It would give you a number of advantages to the background job approach:
- You can easily change the number of concurrent sessions.
- And it will wait until sessions complete before it starts new ones.
- It it easier to abort.
Have a look here for a quick start: https://www.gnu.org/software/parallel/parallel_tutorial.html#A-single-input-source
New contributor
1
Interesting! I'll take a look at this. Do you know if attempting this kind of operation (without the help of Parallel) would risk crashing the hypervisor?
– KuboMD
2 hours ago
add a comment |
Your Answer
StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "106"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});
function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: false,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: null,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});
}
});
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2funix.stackexchange.com%2fquestions%2f516203%2fwhat-happens-if-i-start-too-many-background-jobs%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
2 Answers
2
active
oldest
votes
2 Answers
2
active
oldest
votes
active
oldest
votes
active
oldest
votes
Could all 700 instances possibly run concurrently?
That depends on what you mean by concurrently. If we're being picky, then no, they can't unless you have 700 threads of execution on your system you can utilize (so probably not). Realistically though, yes, they probably can, provided you have enough RAM and/or swap space on the system. UNIX and it's various children are remarkably good at managing huge levels of concurrency, that's part of why they're so popular for large-scale HPC usage.
How far could I get until my server reaches its limit?
This is impossible to answer concretely without a whole lot more info. Pretty much, you need to have enough memory to meet:
- The entire run-time memory requirements of one job, times 700.
- The memory requirements of bash to manage that many jobs (bash is not horrible about this, but the job control isn't exactly memory efficient).
- Any other memory requirements on the system.
Assuming you meet that (again, with only 50GB of RAM, you still ahve to deal with other issues:
- How much CPU time is going to be wasted by bash on job control? Probably not much, but with hundreds of jobs, it could be significant.
- How much network bandwidth is this going to need? Just opening all those connections may swamp your network for a couple of minutes depending on your bandwidth and latency.
- Many other things I probably haven't thought of.
When that limit is reached, will it just wait to begin the next iteration off foo or will the box crash?
It depends on what limit is hit. If it's memory, something will die on the system (more specifically, get killed by the kernel in an attempt to free up memory) or the system itself may crash (it's not unusual to configure systems to intentionally crash when running out of memory). If it's CPU time, it will just keep going without issue, it'll just be impossible to do much else on the system. If it's the network though, you might crash other systems or services.
What you really need here is not to run all the jobs at the same time. Instead, split them into batches, and run all the jobs within a batch at the same time, let them finish, then start the next batch. GUN Parallel can be used for this, but it's less than ideal at that scale in a production environment (if you go with it, don't get too aggressive, like I said, you might swamp the network and affect systems you otherwise would not be touching). I would really recommend looking into a proper network orchestration tool like Ansible (https://www.ansible.com/), as that will not only solve your concurrency issues (Ansible does batching like I mentioned above automatically), but also give you a lot of other useful features to work with (like idempotent execution of tasks, nice status reports, and native integration with a very large number of other tools).
add a comment |
Could all 700 instances possibly run concurrently?
That depends on what you mean by concurrently. If we're being picky, then no, they can't unless you have 700 threads of execution on your system you can utilize (so probably not). Realistically though, yes, they probably can, provided you have enough RAM and/or swap space on the system. UNIX and it's various children are remarkably good at managing huge levels of concurrency, that's part of why they're so popular for large-scale HPC usage.
How far could I get until my server reaches its limit?
This is impossible to answer concretely without a whole lot more info. Pretty much, you need to have enough memory to meet:
- The entire run-time memory requirements of one job, times 700.
- The memory requirements of bash to manage that many jobs (bash is not horrible about this, but the job control isn't exactly memory efficient).
- Any other memory requirements on the system.
Assuming you meet that (again, with only 50GB of RAM, you still ahve to deal with other issues:
- How much CPU time is going to be wasted by bash on job control? Probably not much, but with hundreds of jobs, it could be significant.
- How much network bandwidth is this going to need? Just opening all those connections may swamp your network for a couple of minutes depending on your bandwidth and latency.
- Many other things I probably haven't thought of.
When that limit is reached, will it just wait to begin the next iteration off foo or will the box crash?
It depends on what limit is hit. If it's memory, something will die on the system (more specifically, get killed by the kernel in an attempt to free up memory) or the system itself may crash (it's not unusual to configure systems to intentionally crash when running out of memory). If it's CPU time, it will just keep going without issue, it'll just be impossible to do much else on the system. If it's the network though, you might crash other systems or services.
What you really need here is not to run all the jobs at the same time. Instead, split them into batches, and run all the jobs within a batch at the same time, let them finish, then start the next batch. GUN Parallel can be used for this, but it's less than ideal at that scale in a production environment (if you go with it, don't get too aggressive, like I said, you might swamp the network and affect systems you otherwise would not be touching). I would really recommend looking into a proper network orchestration tool like Ansible (https://www.ansible.com/), as that will not only solve your concurrency issues (Ansible does batching like I mentioned above automatically), but also give you a lot of other useful features to work with (like idempotent execution of tasks, nice status reports, and native integration with a very large number of other tools).
add a comment |
Could all 700 instances possibly run concurrently?
That depends on what you mean by concurrently. If we're being picky, then no, they can't unless you have 700 threads of execution on your system you can utilize (so probably not). Realistically though, yes, they probably can, provided you have enough RAM and/or swap space on the system. UNIX and it's various children are remarkably good at managing huge levels of concurrency, that's part of why they're so popular for large-scale HPC usage.
How far could I get until my server reaches its limit?
This is impossible to answer concretely without a whole lot more info. Pretty much, you need to have enough memory to meet:
- The entire run-time memory requirements of one job, times 700.
- The memory requirements of bash to manage that many jobs (bash is not horrible about this, but the job control isn't exactly memory efficient).
- Any other memory requirements on the system.
Assuming you meet that (again, with only 50GB of RAM, you still ahve to deal with other issues:
- How much CPU time is going to be wasted by bash on job control? Probably not much, but with hundreds of jobs, it could be significant.
- How much network bandwidth is this going to need? Just opening all those connections may swamp your network for a couple of minutes depending on your bandwidth and latency.
- Many other things I probably haven't thought of.
When that limit is reached, will it just wait to begin the next iteration off foo or will the box crash?
It depends on what limit is hit. If it's memory, something will die on the system (more specifically, get killed by the kernel in an attempt to free up memory) or the system itself may crash (it's not unusual to configure systems to intentionally crash when running out of memory). If it's CPU time, it will just keep going without issue, it'll just be impossible to do much else on the system. If it's the network though, you might crash other systems or services.
What you really need here is not to run all the jobs at the same time. Instead, split them into batches, and run all the jobs within a batch at the same time, let them finish, then start the next batch. GUN Parallel can be used for this, but it's less than ideal at that scale in a production environment (if you go with it, don't get too aggressive, like I said, you might swamp the network and affect systems you otherwise would not be touching). I would really recommend looking into a proper network orchestration tool like Ansible (https://www.ansible.com/), as that will not only solve your concurrency issues (Ansible does batching like I mentioned above automatically), but also give you a lot of other useful features to work with (like idempotent execution of tasks, nice status reports, and native integration with a very large number of other tools).
Could all 700 instances possibly run concurrently?
That depends on what you mean by concurrently. If we're being picky, then no, they can't unless you have 700 threads of execution on your system you can utilize (so probably not). Realistically though, yes, they probably can, provided you have enough RAM and/or swap space on the system. UNIX and it's various children are remarkably good at managing huge levels of concurrency, that's part of why they're so popular for large-scale HPC usage.
How far could I get until my server reaches its limit?
This is impossible to answer concretely without a whole lot more info. Pretty much, you need to have enough memory to meet:
- The entire run-time memory requirements of one job, times 700.
- The memory requirements of bash to manage that many jobs (bash is not horrible about this, but the job control isn't exactly memory efficient).
- Any other memory requirements on the system.
Assuming you meet that (again, with only 50GB of RAM, you still ahve to deal with other issues:
- How much CPU time is going to be wasted by bash on job control? Probably not much, but with hundreds of jobs, it could be significant.
- How much network bandwidth is this going to need? Just opening all those connections may swamp your network for a couple of minutes depending on your bandwidth and latency.
- Many other things I probably haven't thought of.
When that limit is reached, will it just wait to begin the next iteration off foo or will the box crash?
It depends on what limit is hit. If it's memory, something will die on the system (more specifically, get killed by the kernel in an attempt to free up memory) or the system itself may crash (it's not unusual to configure systems to intentionally crash when running out of memory). If it's CPU time, it will just keep going without issue, it'll just be impossible to do much else on the system. If it's the network though, you might crash other systems or services.
What you really need here is not to run all the jobs at the same time. Instead, split them into batches, and run all the jobs within a batch at the same time, let them finish, then start the next batch. GUN Parallel can be used for this, but it's less than ideal at that scale in a production environment (if you go with it, don't get too aggressive, like I said, you might swamp the network and affect systems you otherwise would not be touching). I would really recommend looking into a proper network orchestration tool like Ansible (https://www.ansible.com/), as that will not only solve your concurrency issues (Ansible does batching like I mentioned above automatically), but also give you a lot of other useful features to work with (like idempotent execution of tasks, nice status reports, and native integration with a very large number of other tools).
answered 58 mins ago
Austin HemmelgarnAustin Hemmelgarn
6,32111119
6,32111119
add a comment |
add a comment |
It's hard to say specifically how many instances could be run as background jobs in the manner you describe. But a normal server can certainly maintain 700 concurrent connections as long as you do it correctly. Webservers do this all the time.
May I suggest that you use GNU parallel (https://www.gnu.org/software/parallel/) or something similar to accomplish this? It would give you a number of advantages to the background job approach:
- You can easily change the number of concurrent sessions.
- And it will wait until sessions complete before it starts new ones.
- It it easier to abort.
Have a look here for a quick start: https://www.gnu.org/software/parallel/parallel_tutorial.html#A-single-input-source
New contributor
1
Interesting! I'll take a look at this. Do you know if attempting this kind of operation (without the help of Parallel) would risk crashing the hypervisor?
– KuboMD
2 hours ago
add a comment |
It's hard to say specifically how many instances could be run as background jobs in the manner you describe. But a normal server can certainly maintain 700 concurrent connections as long as you do it correctly. Webservers do this all the time.
May I suggest that you use GNU parallel (https://www.gnu.org/software/parallel/) or something similar to accomplish this? It would give you a number of advantages to the background job approach:
- You can easily change the number of concurrent sessions.
- And it will wait until sessions complete before it starts new ones.
- It it easier to abort.
Have a look here for a quick start: https://www.gnu.org/software/parallel/parallel_tutorial.html#A-single-input-source
New contributor
1
Interesting! I'll take a look at this. Do you know if attempting this kind of operation (without the help of Parallel) would risk crashing the hypervisor?
– KuboMD
2 hours ago
add a comment |
It's hard to say specifically how many instances could be run as background jobs in the manner you describe. But a normal server can certainly maintain 700 concurrent connections as long as you do it correctly. Webservers do this all the time.
May I suggest that you use GNU parallel (https://www.gnu.org/software/parallel/) or something similar to accomplish this? It would give you a number of advantages to the background job approach:
- You can easily change the number of concurrent sessions.
- And it will wait until sessions complete before it starts new ones.
- It it easier to abort.
Have a look here for a quick start: https://www.gnu.org/software/parallel/parallel_tutorial.html#A-single-input-source
New contributor
It's hard to say specifically how many instances could be run as background jobs in the manner you describe. But a normal server can certainly maintain 700 concurrent connections as long as you do it correctly. Webservers do this all the time.
May I suggest that you use GNU parallel (https://www.gnu.org/software/parallel/) or something similar to accomplish this? It would give you a number of advantages to the background job approach:
- You can easily change the number of concurrent sessions.
- And it will wait until sessions complete before it starts new ones.
- It it easier to abort.
Have a look here for a quick start: https://www.gnu.org/software/parallel/parallel_tutorial.html#A-single-input-source
New contributor
New contributor
answered 2 hours ago
laenkeiolaenkeio
1065
1065
New contributor
New contributor
1
Interesting! I'll take a look at this. Do you know if attempting this kind of operation (without the help of Parallel) would risk crashing the hypervisor?
– KuboMD
2 hours ago
add a comment |
1
Interesting! I'll take a look at this. Do you know if attempting this kind of operation (without the help of Parallel) would risk crashing the hypervisor?
– KuboMD
2 hours ago
1
1
Interesting! I'll take a look at this. Do you know if attempting this kind of operation (without the help of Parallel) would risk crashing the hypervisor?
– KuboMD
2 hours ago
Interesting! I'll take a look at this. Do you know if attempting this kind of operation (without the help of Parallel) would risk crashing the hypervisor?
– KuboMD
2 hours ago
add a comment |
Thanks for contributing an answer to Unix & Linux Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2funix.stackexchange.com%2fquestions%2f516203%2fwhat-happens-if-i-start-too-many-background-jobs%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
I’m guessing each job uses very little CPU and RAM, is that right?
– Stephen Kitt
3 hours ago
Honestly I have a hard time telling. Htop isn't very helpful - when I'm running one instance the CPU reads:
CPU: 86.9% sys: 13.1% low: 0.0%
and RAM readsMem:3.86G used:178M buffers:2.28G cache:608M
Any guess?– KuboMD
3 hours ago