How should we think about functional units? #8
Replies: 3 comments 3 replies
-
|
I don't know if it's helpful but I've been pondering the idea of a session, which is one user trying to achieve a task. For instance I ask Claude "how do I get this code to compile" and we go back and forth a few times. Or I go to Sora and generate images until I get one that I can use in my LinkedIn post. A great model might answer these perfectly in one turn, or ask a great couple of questions to refine before they do the expensive task of generating code or images. An ok model might take a bunch of turns, or never generate what I want. So while a great model might use more carbon per prompt, it would use far less carbon overall. I think this is important because we want to encourage effective use of AI. So my proposed functional unit is "cost per successful session" or "cost per successful task". This works well for the corporate use case where JP Morgan or somebody updates their customer service chat bot to prompt better and answers 20% more questions without having to call somebody - in this case the denominator goes up by 20% even if carbon per prompt is the same. |
Beta Was this translation helpful? Give feedback.
-
|
WG:
|
Beta Was this translation helpful? Give feedback.
-
|
Maybe to add to this discussion, there are functionnal units defined in the AFNOR general framework on Frugal AI:
It is interesting here to note that retraining is counted separately of initial trainings.
|
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Is a functional unit linked to a "persona", for example a functional unit for the end-user vs. a functional unit for a machine learning engineer/researcher.
This was a discussion topic in the workshops and I thought it was a good way to categorize functional units.
An end-user functional unit might be Per Prompt, since that aligns with both what an end-user understand, but also is a unit of scale that makes sense to them - an end user will be able to scale their usage perhaps just by limiting their prompts so giving them a metric they understand that makes sense.
For a researcher/ml-engineer it might be Per Training Cycle. Since for them what they understand and care about is reducing the carbon emissions for a training cycle.
There was also conversation about how for different functional units you might include different parts of the lifecycle, for instance for an end-user, per-prompt, you might only include the operational costs, e.g. inference and supporting infra of running their prompt and leave everything else out.
Maybe the SCI for AI is a range of functional units which in aggregate cover the whole lifecycle but each functional unit is not responsible for including the whole lifecycle?
Beta Was this translation helpful? Give feedback.
All reactions