qemu-devel
[Top][All Lists]
Advanced

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Qemu-devel] IO accounting overhaul


From: Stefan Hajnoczi
Subject: Re: [Qemu-devel] IO accounting overhaul
Date: Fri, 29 Aug 2014 17:04:46 +0100
User-agent: Mutt/1.5.23 (2014-03-12)

On Thu, Aug 28, 2014 at 04:38:09PM +0200, Benoît Canet wrote:
> I collected some items of a cloud provider wishlist regarding I/O accouting.
> 
> In a cloud I/O accouting can have 3 purpose: billing, helping the customers
> and doing metrology to help the cloud provider seeks hidden costs.
> 
> I'll cover the two former topic in this mail because they are the most 
> important
> business wize.
> 
> 1) prefered place to collect billing IO accounting data:
> --------------------------------------------------------
> For billing purpose the collected data must be as close as possible to what 
> the
> customer would see by using iostats in his vm.
> 
> The first conclusion we can draw is that the choice of collecting IO accouting
> data used for billing in the block devices models is right.

I agree.  When statistics are collected at lower layers it becomes are
for the end user to understand numbers that include hidden costs for
image formats, network protocols, etc.

> 2) what to do with occurences of rare events:
> ---------------------------------------------
> 
> Another point is that QEMU developpers agree that they don't know which policy
> to apply to some I/O accounting events.
> Must QEMU discard invalid I/O write IO or account them as done ?
> Must QEMU count a failed read I/O as done ?
> 
> When discusting this with a cloud provider the following appears: these 
> decisions
> are really specific to each cloud provider and QEMU should not implement them.
> The right thing to do is to add accouting counters to collect these events.
> 
> Moreover these rare events are precious troubleshooting data so it's an 
> additional
> reason not to toss them.

Sounds good, network interface statistics also include error counters.

> 3) list of block I/O accouting metrics wished for billing and helping the 
> customers
> -----------------------------------------------------------------------------------
> 
> Basic I/O accouting data will end up making the customers bills.
> Extra I/O accouting informations would be a precious help for the cloud 
> provider
> to implement a monitoring panel like Amazon Cloudwatch.

One thing to be aware of is that counters inside QEMU cannot be trusted.
If a malicious guest can overwrite memory in QEMU then the counters can
be manipulated.

For most purposes this should be okay.  Just be aware that evil guests
could manipulate their counters if a security hole is found in QEMU.

> Here is the list of counters and statitics I would like to help implement in 
> QEMU.
> 
> This is the most important part of the mail and the one I would like the 
> community
> review the most.
> 
> Once this list is settled I would proceed to implement the required 
> infrastructure
> in QEMU before using it in the device models.
> 
> /* volume of data transfered by the IOs */
> read_bytes
> write_bytes
> 
> /* operation count */
> read_ios
> write_ios
> flush_ios
> 
> /* how many invalid IOs the guest submit */
> invalid_read_ios
> invalid_write_ios
> invalid_flush_ios
> 
> /* how many io error happened */
> read_ios_error
> write_ios_error
> flush_ios_error
> 
> /* account the time passed doing IOs */
> total_read_time
> total_write_time
> total_flush_time
> 
> /* since when the volume is iddle */
> qvolume_iddleness_time

?

> 
> /* the following would compute latecies for slices of 1 seconds then toss the
>  * result and start a new slice. A weighted sumation of the instant latencies
>  * could help to implement this.
>  */
> 1s_read_average_latency
> 1s_write_average_latency
> 1s_flush_average_latency
> 
> /* the former three numbers could be used to further compute a 1 minute slice 
> value */
> 1m_read_average_latency
> 1m_write_average_latency
> 1m_flush_average_latency
> 
> /* the former three numbers could be used to further compute a 1 hours slice 
> value */
> 1h_read_average_latency
> 1h_write_average_latency
> 1h_flush_average_latency
> 
> /* 1 second average number of requests in flight */
> 1s_read_queue_depth
> 1s_write_queue_depth
> 
> /* 1 minute average number of requests in flight */
> 1m_read_queue_depth
> 1m_write_queue_depth
> 
> /* 1 hours average number of requests in flight */
> 1h_read_queue_depth
> 1h_write_queue_depth

I think libvirt captures similar data.  At least virt-manager displays
graphs with similar data (maybe for CPU, memory, or network instead of
disk).

> 4) Making this happen
> -------------------------
> 
> Outscale want to make these IO stat happen and gave me the go to do whatever
> grunt is required to do so.
> That said we could collaborate on some part of the work.

Seems like a nice improvement to the query-blockstats available today.

CCing libvirt for management stack ideas.

Stefan

Attachment: pgp_ZQC8ErUDm.pgp
Description: PGP signature


reply via email to

[Prev in Thread] Current Thread [Next in Thread]