Bug 79353 - Use Unicode 6.1.0 when determining whether an identifier is acceptable or not
Summary: Use Unicode 6.1.0 when determining whether an identifier is acceptable or not
Status: UNCONFIRMED
Alias: None
Product: WebKit
Classification: Unclassified
Component: JavaScriptCore (show other bugs)
Version: 528+ (Nightly build)
Hardware: All All
: P2 Normal
Assignee: Nobody
URL:
Keywords:
Depends on:
Blocks:
 
Reported: 2012-02-23 02:48 PST by Mathias Bynens
Modified: 2012-02-28 09:10 PST (History)
5 users (show)

See Also:


Attachments

Note You need to log in before you can comment on or make changes to this bug.
Description Mathias Bynens 2012-02-23 02:48:31 PST
JavaScript currently uses an older version of the Unicode database. Here are some examples of identifiers that are currently failing because of this, even though they’re valid according to ES 5.1/Unicode 6.1:

* `var \u0cf1;` — http://mothereff.in/js-variables#%5Cu0cf1
* `var \ua7aa;` — http://mothereff.in/js-variables#%5Cua7aa
* `var \u1bba;` — http://mothereff.in/js-variables#%5Cu1bba
* `var a\ua674;` — http://mothereff.in/js-variables#a%5Cua674

Of course, there are many more.

Updating to Unicode 6.1 would improve interoperability.

Is the list of allowed characters in `IdentifierStart` and `IdentifierPart` auto-generated based on a UnicodeData.txt file, or how is this done in JavaScriptCore?