|
|
Subscribe / Log in / New account

What's coming in glibc 2.10

Ulrich Drepper has posted a list of some of the changes which will appear in glibc 2.10. "The new malloc_info function therefore does not export a structure. Instead it exports the information in a self-describing data structure. Nowadays the preferred way to do this is via XML. The format can change over time (it's versioned), some fields will stay the same, other will change. No breakage. The reader just cannot assume that all the information will forever be available in the same form. There is no reader in glibc. This isn't necessary, it's easy enough to write outside glibc using one of the many XML libraries."

to post comments

What's coming in glibc 2.10: XML

Posted Apr 19, 2009 15:40 UTC (Sun) by jkohen (subscriber, #47486) [Link] (19 responses)

Now glibc writes XML? What an over-engineered solution! Even if the burden of using XML will only fall on applications interested in malloc_info, correct XML parsers are not trivial dependencies, and broken parsers shouldn't be used.

Although I don't care much about this particular function, it would be great if simpler and safer alternatives to XML were considered instead, one example being Google's Protocol Buffers.

What's coming in glibc 2.10: XML

Posted Apr 19, 2009 16:19 UTC (Sun) by bvdm (guest, #42755) [Link] (5 responses)

Using XML for this purpose is really not that unreasonable given that very few applications will call this function and they likely already have an XML parser as dependency.

Besides, if you want to export information in a standard and extensible manner, XML is pretty much the way to go. Now Protocol Buffers, that would be over-engineering!

What's coming in glibc 2.10: XML

Posted Apr 19, 2009 21:51 UTC (Sun) by elanthis (guest, #6227) [Link] (2 responses)

It's very unreasonable, because it's a HUGE freaking specification and there's no truly compelling reason not to use a simple list of structs with a corresponding free_mallocinfo() function, just like any number of similar APIs in the libc already use.

What's coming in glibc 2.10: XML

Posted Apr 20, 2009 10:43 UTC (Mon) by alankila (guest, #47141) [Link] (1 responses)

Don't overlook the possibility that this decision is a harbinger for using less binary structs in glibc's API, and more XML. Could happen.

I think it's quite safe to assume that Ulrich means "text with angle brackets" rather than the entirety of XML specification. In other words, even a crude parser is likely to work...

What's coming in glibc 2.10: XML

Posted Apr 20, 2009 15:07 UTC (Mon) by drag (guest, #31333) [Link]

Well ya.. XML is simple. Angle-y brackets and a few other small things then your done.

I don't know why people are getting excited about this. Mallocinfo seems to me to be a very poorly designed part of the POSIX spec and, like the blog entry said, is dependent on the specifics of a particular malloc implimentation.

The particular Mallocinfo in the spec is so widly useless for any sort of modern system that there is just no 'right' way to do it.

So since it's returning statistica data that are dependent on a particular internal inplimentation then the sort of statistics that are relevent to anybody are going to change based on the changes of Malloc.

So instead of requiring application designers to use some sort of custom parser or try to read data whose relevence and meaning, at any point in the future, can and will change he has decided to go with a very commonly used, standard, self-describing format... which is XML.

Makes sense to me.

Who is going to read this information anyway?

Posted Apr 20, 2009 0:27 UTC (Mon) by xoddam (guest, #2322) [Link] (1 responses)

Heap implementation and performance details are not the sort of thing an application can hope to comprehend and adapt to on the fly. I would expect this information to be examined only by developers (often after the fact eg. in logs depicting memory pressure in a production system experiencing performance problems or horrible OOM events), who might prefer a more human-readable representation but shouldn't be afraid of XML in the case of a complex hierarchical heap structure.

Who is going to read this information anyway?

Posted Apr 20, 2009 19:41 UTC (Mon) by jkohen (subscriber, #47486) [Link]

In this case there is really no excuse to go with XML. Even something like /proc/meminfo is more readable to a human than structured markup.

If you want structured data ready for human consumption, you probably need to write a user-space tool, in which case using XML is no better than protocol buffers, JSON and so on; it just has a more complicated spec.

What's coming in glibc 2.10: XML

Posted Apr 19, 2009 20:22 UTC (Sun) by nix (subscriber, #2304) [Link] (11 responses)

Writing XML is easy. It's not as if it needs to parse it. (And tinyxml
*is* a trivial dependency: what is it, 4K?)

I don't like XML much, but this seems like a reasonable application for
it.

What's coming in glibc 2.10: XML

Posted Apr 19, 2009 20:37 UTC (Sun) by jkohen (subscriber, #47486) [Link] (10 responses)

If you plan to only use a subset of XML, then you should to define well what that subset is in advance. Otherwise applications don't know what features they need to support. It doesn't seem like TinyXML supports namespaces for instance, so if a developer uses this library, and glibc starts using namespaces later, the application might break.

XML is way more complex than required for this problem, and there are simpler encoding schemas. Got something against binary formats? JSON could work. Not that I particularly like JSON, but it's simpler than XML and doesn't try to solve more than required by this particular problem.

What's coming in glibc 2.10: XML

Posted Apr 20, 2009 0:33 UTC (Mon) by bvdm (guest, #42755) [Link] (9 responses)

Wowa! Hold on already...

The XML world is a huge alphabet soap, but the XML spec itself (which does *not* include namespaces) is pretty small. Why require a JSON parser if most applications already have an XML dependency?

What's coming in glibc 2.10: XML

Posted Apr 20, 2009 19:26 UTC (Mon) by jordanb (guest, #45668) [Link] (8 responses)

You must live in a different universe than me if you think this is a small spec.

Do you work in aerospace perhaps?

Compare that to the JSON spec, weighing in at all of eight pages. I doubt it'd take more than 100 lines of code for a fully compliant recursive descent parser. And writing an encoder would be just about as easy.

Which makes me wonder if they're *really* including a fully compliant XML encoder into the libc just to spit out a few lines of structured data, or if glibc is going to be joining the long line of shitty systems that produce broken XML by concatenating strings.

What's coming in glibc 2.10: XML

Posted Apr 20, 2009 22:42 UTC (Mon) by nix (subscriber, #2304) [Link]

Of *course* it's going to produce XML by concatenating strings, but I see
no reason why it has to be 'broken'. It's not as if it's going to be doing
fully general-purpose data->XML transformation: its emitting *one
structure* with contents fully controlled by the libc.

What's coming in glibc 2.10: XML

Posted Apr 21, 2009 4:24 UTC (Tue) by bvdm (guest, #42755) [Link] (6 responses)

Have you read it? It is wonderfully complete and quite educational. If the JSON spec was as well written it would probably be about 3/4 the length.

What's coming in glibc 2.10: XML

Posted Apr 21, 2009 21:56 UTC (Tue) by jordanb (guest, #45668) [Link] (5 responses)

I haven't read it in its entirety. It does seem to be well written, I'll agree to that.

The JSON spec is terse but it also rigorously defines the syntax of a very simple data serialization format. There are no omissions that make it incomplete. This is the advantage of JSON for simple data serialization tasks.

What's coming in glibc 2.10: XML

Posted Apr 22, 2009 4:15 UTC (Wed) by dlang (guest, #313) [Link] (4 responses)

the JSON spec is defining a subset of javascript. does it really fully define everything in it? or does it refer you to the main javascript definitions for the gory details and corner cases?

What's coming in glibc 2.10: XML

Posted Apr 22, 2009 16:09 UTC (Wed) by jordanb (guest, #45668) [Link] (3 responses)

JSON was designed to be a pure subset of Javascript so they could
market it for web development as a way to send data out to your
client-side JS by simply serializing JSON objects into <script>
tags in the webpage. That's the only reason for the focus on
Javascript -- not to rely on the ECMAScript standard. In fact the
only reference to ECMAScript is informational:

"JavaScript Object Notation (JSON) is a text format for the
serialization of structured data. It is derived from the object
literals of JavaScript, as defined in the ECMAScript Programming
Language Standard, Third Edition [ECMA]."

The standard is so small because it only has six data
types ('object' (essentially associative array), list, number,
boolean, string, and 'null'). It has no ability to do node
attributes and is only tree-based to the extent that objects and
arrays can contain more of them. It is certianly not a rival to
XML for dealing with large scale or very complicated or nuanced
data. But it is an excellent alternative when you have small
amounts of structured data that you wish to serialize.

Another advantage of JSON is that if you have simple mostly
tabular datasets they can be serialized (with judicious use of
whitespace) in a manner that's both machine and
human-readable. For fun and as an example I decided to 'encode'
my /proc/cpuinfo in JSON:

[{
processor : 0,
vendor_id : "GenuineIntel",
cpu_family : "6",
model : "8",
model_name : "Pentium III (Coppermine)",
stepping : 1,
cpu_MHz : 498.283,
cache_size : [ 262144, "256 KB" ],
fdiv_bug : false,
hlt_bug : false,
f00f_bug : false,
coma_bug : false,
fpu : true,
fpu_exception : true,
cpuid_level : 2,
wp : true,
flags : [ "fpu", "vme", "de", "pse", "tsc", "msr", "pae", "mce", "cx8",
"sep", "mtrr", "pge", "mca", "cmov", "pat", "pse36", "mmx", "fxsr", "sse", "up" ],
bogomips : 997.69,
clflush_size : 32
}]

What's coming in glibc 2.10: XML

Posted Apr 22, 2009 16:11 UTC (Wed) by jordanb (guest, #45668) [Link]

Of course that looked nicer before lwn nuked my whitespace but you get the idea.

What's coming in glibc 2.10: XML

Posted Apr 22, 2009 18:51 UTC (Wed) by bronson (subscriber, #4806) [Link] (1 responses)

This is actually invalid JSON. You need to quote the key names:

"hlt_bug" : false,
"f00f_bug" : false,
"coma_bug" : false,

Even so, I'd still much rather work with JSON than XML!

What's coming in glibc 2.10: XML

Posted Apr 22, 2009 18:56 UTC (Wed) by jordanb (guest, #45668) [Link]

Heh. I was cringing when I submitted that because I was sure I'd make a mistake. I thinking too hard about making sure there were commas were all in order. :P

What's coming in glibc 2.10: XML

Posted Apr 20, 2009 15:33 UTC (Mon) by iabervon (subscriber, #722) [Link]

I think it's weirder that glibc will be outputting a big string than that the string will be XML. What I think would be more sensible is if it output a data structure that looked like something you might get by parsing XML (ignoring comments, irrelevant whitespace, details of how characters were written, etc). That is, it would be a struct with a name, an array of attribute keys and values, an array of children, and a string for character data.

This would have the same extensibility benefits as XML, but avoid most of the parsing overhead. (Of course, the main practical extensibility benefit comes from using ASCII decimal strings of arbitrary length for numbers, which then requires a bit of parsing.)

What's coming in glibc 2.10

Posted Apr 19, 2009 16:54 UTC (Sun) by welinder (guest, #4699) [Link] (3 responses)

mallinfo is the way to monitor how much memory an application is using.
Having to go through XML for that sounds like serious overkill. That
said, there are probably other pieces of information from malloc that
it actually makes sense to provide in xml form.

What's coming in glibc 2.10

Posted Apr 19, 2009 17:40 UTC (Sun) by eklitzke (subscriber, #36426) [Link] (2 responses)

Actually you can read the memory usage for a process from /proc (e.g. /proc/PID/statm for a machine readable form). The information given by mallinfo is more detailed stuff about what the memory allocator is actually doing internally, not how much memory your process is using.

What's coming in glibc 2.10

Posted Apr 19, 2009 22:30 UTC (Sun) by oak (guest, #2786) [Link] (1 responses)

Both are needed. Mallinfo tells whether the process is fragmenting its
heap. Checking that may be a performance issue, so needing to parse XML
sounds strange. And if Glibc does allocs for the XML output, it would be
even stranger...

What's coming in glibc 2.10

Posted Apr 19, 2009 23:50 UTC (Sun) by foom (subscriber, #14868) [Link]

Indeed, it seems to me that a sysctl-like API would make more sense.

What's coming in glibc 2.10

Posted Apr 19, 2009 17:53 UTC (Sun) by ikm (guest, #493) [Link] (1 responses)

Jon's trolling, hehe :)

corbet trolling

Posted Apr 21, 2009 0:01 UTC (Tue) by xoddam (guest, #2322) [Link]

Corbet is indeed trolling and he's got us all going ...

But when is Drepper going to come down amongst us in person and tell us IT ISN'T SO?

What's coming in glibc 2.10

Posted Apr 19, 2009 20:23 UTC (Sun) by nix (subscriber, #2304) [Link] (3 responses)

Am I the only person who thinks STT_GNU_IFUNC sounds seriously cool? It's
like having a private per-symbol hook into the dynamic linker :)

What's coming in glibc 2.10

Posted Apr 19, 2009 21:50 UTC (Sun) by elanthis (guest, #6227) [Link] (1 responses)

Which makes me wonder if there are any potential security issues with it. Would this, for example, defeat SELinux protections by allowing an application to inject executable code into the context of the linker (assuming it's a separate context from the application itself) ?

What's coming in glibc 2.10

Posted Apr 19, 2009 22:15 UTC (Sun) by nix (subscriber, #2304) [Link]

It isn't a separate context. It's more like injecting code into the
implicit first call into the dynamic linker that happens the first time
any given symbol is referenced[1]. The security context is unchanged: all
the dynamic linker normally does is relocate that PLT entry. It's just
that now you can have it pick *what* to relocate first. It's no more done
in a different context than is calling a function through a function
pointer: that's all it really is.

(hm, and it looks like it might become STT_IFUNC in time and not be
GNU-specific after all. Even better.)

[1] an oversimplification: actually that implicit call happens the first
time a PLT entry is jumped to, but that's nearly the same thing...

What's coming in glibc 2.10

Posted Apr 19, 2009 22:55 UTC (Sun) by jreiser (subscriber, #11027) [Link]

... a function pointer to a function that takes no argument and returns the real function pointer to use ...

It would be useful to have some parameters: a pointer to the symbol table entry for the symbol in question (and the string table, ...), and a pointer to the dynamic linker context (search path: chain of struct link_map *) that applies. Then the same intercepting function could handle multiple names (there are many str* and mem* functions, for instance) and might be able to handle better some special cases such as compatibility matching, etc.

Why the vehement objections to decimal floating-point?

Posted Apr 20, 2009 0:25 UTC (Mon) by stevenj (guest, #421) [Link] (14 responses)

Certain special interest groups subverted the standardization process (again) and pressed through changes to introduce in the C programming language extensions to support decimal floating point computations. 99.99% of all the people will never use this stuff and still we have to live with it.

My understanding was that decimal floating point is actually extraordinarily useful in circumstances like banking and accounting where human inputs (which are invariably in decimal) have to be preserved exactly, but the range of fixed-point representations is too narrow. (And good arguments have been made by various people with long experience in floating-point arithmetic, such as Kahan, that decimal floating point's elimination of rounding in binary-decimal conversion will eliminate a lot of silly errors in other fields too, and hence should become the norm as long as the speed is adequate.) And the latest revision of IEEE-754 describes an efficient way to encode decimal floating point into binary, with the possibility of future hardware support as well, so C is being revised to support this capability.

So why is Drepper digging in his heels?

Why the vehement objections to decimal floating-point?

Posted Apr 20, 2009 0:43 UTC (Mon) by tbrownaw (guest, #45457) [Link] (7 responses)

My understanding was that decimal floating point is actually extraordinarily useful in circumstances like banking and accounting where human inputs (which are invariably in decimal) have to be preserved exactly, but the range of fixed-point representations is too narrow.

I thought that with money you always wanted high-precision fixed-point with hard errors on numeric overflow, such as "15 digits with two of them after the decimal point". Floating point anything would mean that when you got enough dollars your cents would start getting rounded off, so what I understand is typically done is to use rationals with a fixed denominator of 100 (or whatever your money system uses).

Why the vehement objections to decimal floating-point?

Posted Apr 20, 2009 1:47 UTC (Mon) by ringerc (subscriber, #3071) [Link] (6 responses)

Generic rationals can be a screaming nightmare to work with. Rationals are OK if you use a fixed denominator for the type and are really careful about the sorts of calculations you perform and the ordering of those calculations. They're still a pain, just not as bad as variable-denominator rationals.

It seems to generally be quite sufficient to use double precision floating-point for this sort of thing. You just have to ensure you're running in strict IEEE floating point mode, trap floating point exceptions, and allow enough precision that you have at least a couple of significant figures of breathing room.

I've been fairly happy with PostgreSQL's `DECIMAL' type. It's a fixed-precision base-10 decimal that's REALLY nice to work with:

test=# SELECT (DECIMAL '100000000000000000000.0' + DECIMAL '0.0000000000100000000000000001') * DECIMAL '4.0' AS result; 
                       result                        
-----------------------------------------------------
 400000000000000000000.00000000004000000000000000040

test=# SELECT DECIMAL '400000000000000000000.0000000000000000000000000000000000000000000000000000040' / DECIMAL '1.1' AS result2;
                                    result2                                    
-------------------------------------------------------------------------------
 363636363636363636363.6363636363636363636363636363636363636363636363636363673
(1 row)

Having something like this in the core C language specification would be quite delightful.

(Of course, on a random and unrelated rant: a proper unicode string type would be a HUGE improvement to C/C++. The poorly-specified, painful to work with wchar_t with its variation across platforms and implementations, and its total lack of associated encoding conversion functions doesn't really do the job. Let's not even talk about std::wstring .)

Anyway ... I'd actually be really interested in some good references on numeric type choice and proper calculation methods in financial applications.

wchar_t

Posted Apr 20, 2009 8:37 UTC (Mon) by tialaramex (subscriber, #21167) [Link] (5 responses)

You don't need a Unicode string type in C, and it's probably a mistake to ask for one to be built into C++ (but I can't tell, maybe C++ standardisation is about trying to /collect/ mistakes at this point).

wchar_t is a legacy of the mistaken belief that Unicode was (as some documents from a decade or more ago declared) the encoding of all world symbols into a 16-bit value. Once UCS-2 was obsolete, wchar_t was obsolete too, don't use it. Use UTF-8 on the wire, on disk and even in memory except when you're doing heavyweight character processing, and then use UTF-32, ie uint32_t or at a pinch (since the tops bits are unused anyway) int.

The only real non-legacy argument for UTF-16 was that it's less bytes than UTF-8 for texts in some writing systems, notably Chinese. But the evidence of the last couple of decades is that the alphabetic and syllabic writing systems will eat the others alive, the majority of the world's population may yet be speaking Chinese in our lifetimes, but if so they'll write it mostly in Roman script destroying UTF-16's size advantage.

wchar_t

Posted Apr 20, 2009 13:33 UTC (Mon) by mrshiny (guest, #4266) [Link] (3 responses)

Chinese has only about 410 unique syllables (2050 if you include tones). There are thousands of words and many sentences which, even if tones are properly conveyed, are ambiguous. I would be surprised if the current romanizations replaced Chinese characters. I would be less surprised to see an alphabet arise instead, much like what the Japanese have.

wchar_t

Posted Apr 20, 2009 18:55 UTC (Mon) by proski (subscriber, #104) [Link] (2 responses)

It's already happening in Taiwan: http://en.wikipedia.org/wiki/Bopomofo

wchar_t

Posted Apr 21, 2009 1:19 UTC (Tue) by xoddam (guest, #2322) [Link] (1 responses)

On the other hand, pinyin is pretty much universal on the mainland, often as a way of specifying regional or personal pronunciation.

Vietnam has successfully switched (almost) entirely from han tu to the Latin alphabet, albeit with a forest of diacritics. Chinese might one day do the same, but it's unlikely since there is such a variety of Chinese speech and usage. Unlike Vietnamese, Chinese national identity has never been a matter of shared pronunciation.

Both pinyin and bopomofo have some chance of evolving to make it possible to write both pronunciation and semantics reproducibly in the same representation, but neither is likely to become a universal replacement for hanzi, since they lose the advantage (not meaningfully damaged by the Simplified/Traditional split) that the several very different Chinese languages become mutually intelligible when written down.

Universal alphabetisation of Chinese won't be possible until the regional differences become better acknowledged, so people learn literacy both in their first dialect and in the "standard" language(s).

As for the relatively low count of "unique" symbols -- the whole idea of unifying hanzi and the Japanese and Korean versions using the common semantics to reduce the required code space and "assist" translations and text searches has met great resistance, especially in Japan, and despite it there are now nearly 100,000 distinct characters defined in Unicode. 16 bits was always a pipe dream.

It is ultimately necessary (ie. required by users) to represent distinct glyphs uniquely; Unicode still doesn't satisfy many users precisely because it tries not to have too many distinct code points; probably it never will.

I expect one day the idea of choosing a font based on national context will be abandoned, and the code point count will finally explode, defining one Unicode character per glyph.

wchar_t

Posted Apr 30, 2009 17:27 UTC (Thu) by pixelpapst (guest, #55301) [Link]

> I expect one day the idea of choosing a font based on national context will be abandoned, and the code point count will finally explode, defining one Unicode character per glyph.

I agree. And I think when this happens, we just *might* see a revival of UTF-16 in Asia - in a modal form. So you wouldn't repeat the high-order surrogate when it is the same as that of the previous non-BMP character.

This would pack these texts a bit tighter than UTF-8 or UCS-4 (can encode 10 bits per low-order surrogate), while being a bit easier to parse than all the Escape-Sequence modal encodings.

IMHO, let's see.

wchar_t

Posted Apr 21, 2009 0:25 UTC (Tue) by xoddam (guest, #2322) [Link]

wchar_t is 32-bit by default in g++ and in the stddef.h usually used by gcc.

There is a g++ compiler option -fshort-wchart to change the intrinsic type in C++, and you can use alternative headers or pre-define "-D__WCHAR_T__=uint16_t" for C, but this is pretty unusual on Linux except when cross-compiling for another platform (or building WINE).

Why the vehement objections to decimal floating-point?

Posted Apr 20, 2009 1:01 UTC (Mon) by mgb (guest, #3226) [Link] (5 responses)

64-bit integers are adequate to deal with most banking and financial calculations today. Currencies are generally denoted in cents or mills, although the Zimbabwean dollar might more appropriately be denoted in billions.

128-bit floats can be used to store very large integers today and pure 128-bit integers are waiting in the wings.

Decimal floats are symptomatic of poor design.

Why the vehement objections to decimal floating-point?

Posted Apr 20, 2009 8:25 UTC (Mon) by epa (subscriber, #39769) [Link] (4 responses)

64-bit integers are adequate to deal with most banking and financial calculations today.
Not all numbers used in finance are integers. Consider exchange rates and interest rates, for a start. If you were particularly perverse you could decide to use 64-bit ints for everything, with some way of encoding the number of decimal places (or binary places), but in that case you have effectively reinvented a floating point math library.
Decimal floats are symptomatic of poor design.
Not at all. They are often the best match to what the user and the rest of the world requires. It is accepted that 1/3 gives a recurring decimal .333... but no accountant wants their computer system to introduce rounding errors, no matter how minute, when calculating 1/5 (which is .0011... in binary). Or do you mean that *floating* point decimal is a bad idea, and it's better to use fixed point with a certain fixed number of digits precision? There is certainly a case for that.

Why the vehement objections to decimal floating-point?

Posted Apr 20, 2009 16:31 UTC (Mon) by stevenj (guest, #421) [Link] (3 responses)

A lot of people here are proposing that decimal fixed point is just as good or better than decimal floats.

I'm a little skeptical of this, based on my experience with scientific computation: there are many, many circumstances when both the input and output of the computation appear to be in a range suitable for fixed-point representation, but the intermediate calculations will have vastly greater rounding errors in fixed point than in floating point. And fixed-point error analysis in the presence of rounding and overflow is a nightmare compared to floating point.

Decimal floating point gives you the best of both worlds. If the result of each calculation is exactly representable, it will give you the exact result. (Please don't raise the old myth that floating-point calculations add some mysterious random noise to each calculation!) There is no rounding when decimal inputs are entered, so human input is preserved exactly. And if the result is not exactly representable, its rounding characteristics will be much, much better than fixed point. (And don't try to claim that financial calculations never have to round.)

Note that the IEEE double-precision (64-bit) decimal-float format has a 16 decimal-digit significand (and there is also a quad-precision decimal float with a 34 decimal-digit significand). I would take this over 64-bit fixed point any day: only nine bits of this are sacrificed in IEEE to give you a floating decimal point and fixed relative precision over a wide dynamic range.

Why the vehement objections to decimal floating-point?

Posted Apr 20, 2009 16:34 UTC (Mon) by stevenj (guest, #421) [Link] (2 responses)

(By "result of each calculation is exactly representable" I am of course including intermediate calculations. Note that this is equally true in fixed-point and integer arithmetic.)

Why the vehement objections to decimal floating-point?

Posted Apr 25, 2009 12:29 UTC (Sat) by dmag (guest, #17775) [Link] (1 responses)

Fixed-point won't loose information on simple calculations, but there is a possibility some intermediate results will saturate your representation. For example, if you square a number, add 1 and take the square root. For large numbers, the square isn't likely to be representable.

Floating point has the opposite problem. The intermediate calculations won't blow up, but you can lose precision even in simple cases.

Most people don't have a correct mental model of floating point. Floating point has a reputation for being 'lossy' because it can loose information in non-obvious ways.

$ irb
>> 0.1 * 0.1 - 0.01
=> 1.73472347597681e-18
Sometimes the answer is to store in fixed point, but calculate in floating point (and do appropriate rounding during conversion back to fixed).

Why the vehement objections to decimal floating-point?

Posted Apr 18, 2011 22:37 UTC (Mon) by stevenj (guest, #421) [Link]

Your example makes no sense; the result would be computed exactly in decimal floating point.

More generally, in essentially any case where decimal fixed point with N digits would produce exact results, decimal floating point with an N-digit significand would also produce exact results. The only sacrifice in going from fixed to (decimal) floating point is that you lose a few bits of precision to store the exponent, and in exchange you get vastly bigger dynamic range and much more sensible roundoff characteristics.

You're certainly right that many people don't have a correct mental model of floating point, however.

What's coming in glibc 2.10

Posted Apr 20, 2009 4:31 UTC (Mon) by xorbe (guest, #3165) [Link] (1 responses)

I was hoping that glibc would upload the data to the localhost MySQL server at runtime, and then my program could query the results back...

What's coming in glibc 2.10

Posted Apr 20, 2009 7:37 UTC (Mon) by bvdm (guest, #42755) [Link]

No, I think a REST-style Web 2.0 enabled solution will work better..

LOL

What's coming in glibc 2.10

Posted Apr 20, 2009 12:13 UTC (Mon) by oseemann (guest, #6687) [Link] (2 responses)

Are there any other libraries that return a buffer with XML data to the caller instead of a C struct?

What's coming in glibc 2.10

Posted Apr 20, 2009 19:36 UTC (Mon) by jkohen (subscriber, #47486) [Link]

Any library intended to write data in XML form, naturally. You can however find much more creative uses of XML at the TheDailyWTF.com. Sadly, now I wouldn't be surprised to see glibc popping up there soon.

What's coming in glibc 2.10

Posted Apr 20, 2009 22:25 UTC (Mon) by nix (subscriber, #2304) [Link]

Are there *any* libraries or protocols with similar binary compatibility
guarantees? (Other than X, whicih is having similar problems: there's no
way *it* can use XML over the wire, but it's recently had to
introduce 'generic events' because it's run out of spare events...)


Copyright © 2009, Eklektix, Inc.
Comments and public postings are copyrighted by their creators.
Linux is a registered trademark of Linus Torvalds