select only involves polling when done incorrectly.
And that can consume 60% to 70% of a cpu just polling.
Yeah, when I use select, it doesn't burn CPU when waiting.
| [reply] |
Well, select (and poll) do have inherent limitations. For each call to select the kernel must check all specified descriptors to see if they are ready. Performance of that depends on the number of descriptors. Also, in a busy application select can be called very often. I think it's fair to call it 'polling'.
With epoll and other similar mechanisms, it's only necessary to register all 'interesting' descriptors once, and then, when IO happens on some descriptor, the kernel checks if the application is interested in it. Therefore, performance is determined by the number of IO events.
If we have a ton of descriptors, but, at any given time, IO actually happens only on some small percent of them, epoll will vastly outperform select, because, yes, it does less 'polling'.
Yeah, when I use select, it doesn't burn CPU when waiting.
But do you have 10k concurrent connections :)
| [reply] [d/l] [select] |
| [reply] |
Not a good example, select and poll are indeed slow and not recommended for "a few hundren tcp clients"; use epoll/kqueue/signal-driven IO (well, maybe not signal-driven IO). | [reply] [d/l] [select] |
| [reply] |
Hey, BrowersUK, some of your objections are just bizarre.
They're not available from Perl.
What
"epoll doesn't work with timers, signals... yes, there are the calls signalfd(), and timerfd_create()... they don't help much"
??? BTW, none of the IO multiplexers (to my knowledge) do anything useful with hard disk files.
"epoll maybe O(1) for event notifications, but it remains O(n) for modifications to the interest set"
you found that a problem in practice? Tell me more...
fast changing, dynamic descriptor sets require many and calls into the kernel for epoll_ctl(), which gets expensive.
"fast" compared to what? compared to the speed of CPUs, or compared to the speed of IO that you're presumably doing on these descriptors in between opening and closing them?
kqueue addresses most of that, but retains one major flaw in modern systems: Per process interest sets.
Now, I haven't actually used kqueue (since that's a BSD thing). But you said "retains", and that's wrong: you can create as many epoll instances (sets) as you like, and monitor them separately (using threads, if you like). Actually, kqueue man page says:
The kqueue() system call fails if:
[ENOMEM] The kernel failed to allocate enough memory for the k
+ernel queue.
[EMFILE] The per-process descriptor table is full.
[ENFILE] The system file table is full.
Where does it say "you already have one queue in this process and can't get more"?
Basically, it hasn't caught up with the advent of the multi-core architectures that are now ubiquitous; which means you can't easily split your loads across those multiple cores by using multiple thread running separate event loops.
Sure you can, if that's what you want to do, for some reason...
Really, your points look pretty strange to me, except maybe for adding more than one fd at a time to epoll instance. Well, that looks like it wouldn't be difficult to implement, but it wasn't, so I guess other people didn't feel the need for it either. | [reply] [d/l] [select] |