tobiastomit’s not really about my use case. I can for sure build some search, or reuse yours, but still: I think the goal of the JSON format would be to have the data available in a easy to use way.
tobiastomif you requite some tools, or some own implementation, to get easy stuff done, like get all h-cards, I think it makes it much harder for everyone.
tobiastommy use case is not really related to that, I’m just trying to get a summary for an URL right now, which will support microformats, og and others, but this just came to my mind while implementing it.
barnabywaltersso the canonical JSON is a first step towards that, allowing us to make parsers interoperable, and build tools which enhance that canonical data
barnabywaltersone of the things I’ll be extracting from my feed reader work into it’s own package is a toolkit which will do as much of this normalizing work as possible
tobiastomyeah, I see your valid point about correcting stuff and so on, but I think having an optimized thing, and not needing an optimizer in the first place would be a better start.
barnabywalterstobiastom: in theory that’s correct, however it requires the parsers to try to understand the semantics of what they’re parsing, not just the structure
barnabywaltersbut it’s necessary a) as a way of deterministically comparing and testing parsers, and b) as a standard representation on top of which better tools can be created
tobiastomagreed on the parser developer point. but I think the main question is: do we want to keep the space for optimizers open on the future, or should the parsers generate that in the first place.
tobiastomif I would like to implement a parser now, I would be required to create the JSON structure now, and then the optimized version. That’s more work for me, so why do it?
tobiastomYeah, I think you are right. My intention is just to generic right now. For me the best case would be if all the different solutions (like og, rdf, whatever) would sit on a table an talk about the optimized format, but I assume that will not happen. :)
barnabywaltersalso, it’s actually *less* work over all to have the canonical “raw” JSON format, because then parser developers only need to worry about making that, and other people can work on normalizers — and one normalizer will work with any parser
barnabywalterstobiastom: it’s understandable. I’ve spent a *lot* of time yak shaving on these sort of enabling projects and am just now getting to the stage where I can come up with an idea for a project, and start working on business logic right away
barnabywalterstobiastom: I suspect your suite is going to be objectively easier to use, so once it’s in use by a couple of parsers (certainly php-mf2, point the mf2py devs at it as well) if it works then you can suggest making it the canonical test suite
tantekbarnabywalters: you'll have to bring that up with glennjones - I think he has an explanation about why the tests marked up with microformats are a good thing
barnabywalterstantek: AFAIK currently tobiastom’s tests are generated from glennjones’s tests (as they are impressively thorough), so the two could coexist
barnabywaltersbut if one is significantly easier both to maintain and for developers to use, as I suspect is the case, then IMO that should be the test suite we point parser developers towards