Re: performance for high-volume log insertion - Mailing list pgsql-performance
From | James Mansion |
---|---|
Subject | Re: performance for high-volume log insertion |
Date | |
Msg-id | 49EF83AF.2090307@mansionfamily.plus.com Whole thread Raw |
In response to | Re: performance for high-volume log insertion (Stephen Frost <sfrost@snowman.net>) |
Responses |
Re: performance for high-volume log insertion
Re: performance for high-volume log insertion |
List | pgsql-performance |
Stephen Frost wrote: > You're re-hashing things I've already said. The big win is batching the > inserts, however that's done, into fewer transactions. Sure, multi-row > inserts could be used to do that, but so could dropping begin/commits in > right now which probably takes even less effort. > Well, I think you are seriously underestimating the cost of the round-trip compared to all the other effects (possibly bar the commits). When I tested the union insert technique on SQLServer and Sybase I got measurable improvements going from 100 row statements to 200 row statements, though I suspect in that case the per-statement overheads are quite high. I expected improvements from 10 to 20 row batches, but it carried on getting better for a long time after that. The Sybase parser runs out of workspace first. > No, as was pointed out previously already, you really just need 2. A > And I'm disagreeing with that. Single row is a given, but I think you'll find it pays to have one round trip if at all possible and invoking multiple prepared statements can work against this. > see if there's really much of a performance difference between a > 50-insert prepared statement, and 50 1-insert prepared statements. If > they're both done in larger transactions, I don't know that there's > really alot of performance difference. > I think you'll be surprised, but the only way is to test it. And also the simple 50 row single insert as text. See if you can measure the difference between that and the prepared statement. > storage overhead? indexing overhead? We're talking about prepared > statements here, what additional storage requirement do you think those > would impose? What additional indexing overhead? I don't believe we > actually do anything differently between prepared statements and > multi-row inserts that would change either of those. > That's my point. You will brickwall on the actual database operations for execution early enough that the efficiency difference between parse-and-execute and prepared statements will be hard to measure - at least if you have multi-row statements. But this really needs testing and timing. > Ah, latency is a reasonable thing to bring up. Of course, if you want > to talk about latency then you get to consider that multi-insert SQL > will inherently have larger packet sizes which could cause them to be > delayed in some QoS arrangements. > No, I mean latency from round trips from the client to the server process. I don't know why you think I'd mean that. > As I said, most of this is a re-hash of things already said. The > low-hanging fruit here is doing multiple inserts inside of a > transaction, rather than 1 insert per transaction. Regardless of how > that's done, it's going to give the best bang-for-buck. It will > complicate the client code some, regardless of how it's implemented, so > that failures are handled gracefully (if that's something you care about > anyway), but as there exists some queueing mechanisms in rsyslog > already, hopefully it won't be too bad. > I think you have largely missed the point. There are two things here: 1) how many rows per commit 2) how many rows per logical RPC (ie round trip) between the client and server processes We are agreed that the first is a Very Big Deal, but you seem resistant to the idea that the second of these is a big deal once you've dealt with the former. My experience has been that its much more important than any benefits of preparing statements etc, particularly if the use of a prepared statement can make it harder to do multi-row RPCs because the protocol doesn't allow pipelining (at least without things getting very hairy). Clearly 'copy' is your friend for this too, at least potentially (even if it means streaming to a staging table). James
pgsql-performance by date: