Looks like you're trying to pull information from your wireless 802.11b AP.
The data is contained in base64 encoded byte arrays which are simply the data points concatenated together (fixed-length record format).
The JS code to split them out is in the bridge/xxxx.html pages (each page parses it's own record format).
Depending on which value you want, you should be able to pull the data using Javascript of your own and parse it using their code directly.
You can execute Javascript directly from the command-line using one of several JS interpreters.
"apt-get install spidermonkey-bin" in Ubuntu should get what you need, although it's not necessarily the best.
Once installed, the command to run the js interpreter is "js".
You can even use it as a scripting language with "#!/usr/bin/env js" as the first line of any executable text file containing Javascript commands.
Not sure if that helps enough, the code to pull the data from the website might be a bit tough (asyncHttpRequest is a browser function, not part of Javascript), but perhaps a combination of wget with the command-line Javascript engine will get what you need.
kitepilot@kitepilot.com wrote:
> Hello geeky World...
>
> I have a bunch of Senao CB3 radios in my network which will only take a WEB
> interface for configuration, and such interface uses javascript to parse a
> binary record of unknown structure at load time to display the content of
> the configuration parameters.
>
> You can see a dump as:
> wget --recursive --page-requisites --convert-links --debug --user=XXXXXXXX
> --password=XXXXXXXX 172.28.0.52
> at:
> http://www.kitepilot.com/radio.tar.bz2
>
> Now, this is the challenge:
> I want to extract information from the page without human intervention.
>
> Before the page is displayed, my information is in the binary record, and
> other than reverse engineering the format, it can only be extracted using
> javascript.
>
> This is what I have tried:
> I dumped the page with lynx, which doesn't work cuz lynx ignores javascript.
> I dumped the page with wget, which leaves me with all I need, but then I
> can't find a browser that extracts my info and I am still in the same
> boat...
> Neither lynx, w3b or links will work, for different reasons each one.
>
> Sooo...
> I decided to write a perl script, but because my perl skills are rusted
> beyond WD40 hope, it's becoming a stinking project... :(
>
> What I am trying to do is simple:
> There is a WEB site that requires authorization, that uses javascript to
> extract the information I need, and I want to parse it out.
> There's got to be a simple way. :(
> I ran out of ideas...
> Any one?
> Thanks!
> Enrique
> ---------------------------------------------------
> PLUG-discuss mailing list - PLUG-discuss@lists.plug.phoenix.az.us
> To subscribe, unsubscribe, or to change your mail settings:
> http://lists.PLUG.phoenix.az.us/mailman/listinfo/plug-discuss
>
---------------------------------------------------
PLUG-discuss mailing list -
PLUG-discuss@lists.plug.phoenix.az.us
To subscribe, unsubscribe, or to change your mail settings:
http://lists.PLUG.phoenix.az.us/mailman/listinfo/plug-discuss