theprojecter / whazzup Goto Github PK
View Code? Open in Web Editor NEWAutomatically exported from code.google.com/p/whazzup
Automatically exported from code.google.com/p/whazzup
What it says.
Original issue reported on code.google.com by [email protected]
on 24 Jul 2011 at 10:39
The tokenizer will sometimes create tokens like "berlusconi'", which isn't
right. Need to also look for other bad tokenizations.
Original issue reported on code.google.com by [email protected]
on 22 Feb 2011 at 8:47
It should be possible to add sites simply by uploading an OPML file into
Whazzup.
Original issue reported on code.google.com by [email protected]
on 16 Feb 2011 at 1:52
This should keep the impact of checking feeds fairly low.
Original issue reported on code.google.com by [email protected]
on 1 Aug 2011 at 7:56
In the item report, that is.
Original issue reported on code.google.com by [email protected]
on 3 Aug 2011 at 12:30
We don't need to recheck bad feeds every time we check feeds.
Original issue reported on code.google.com by [email protected]
on 28 Jul 2011 at 12:12
For one thing, need to flush the log file.
Original issue reported on code.google.com by [email protected]
on 24 Jul 2011 at 10:41
We need to break up the bigger tasks more so they don't run all at once and hog
the queue.
Original issue reported on code.google.com by [email protected]
on 24 Jul 2011 at 10:41
u'\u0441\u0431, 13 feb 2010 15:51:30 +0300'
This from a Russian blog.
Original issue reported on code.google.com by [email protected]
on 26 Jul 2011 at 8:17
Now we rerun the scheduled tasks every time the queue is restarted. Not good.
Instead, move scheduling out to actual cron, and use send.py to start the tasks.
Original issue reported on code.google.com by [email protected]
on 23 Jul 2011 at 12:30
What it says.
Original issue reported on code.google.com by [email protected]
on 20 Aug 2011 at 6:06
Use send.py to inject the tasks in the queue.
Original issue reported on code.google.com by [email protected]
on 24 Jul 2011 at 10:42
This is one of the biggest obstacles to scaling. Need to analyze it to see
where it actually spends the time. If it turns out that parsing the text is the
time-waster then we should consider caching the word vector, for example by
marshalling the hash into a database field.
Original issue reported on code.google.com by [email protected]
on 26 Jul 2011 at 7:34
That is, feeds which haven't been updated in a long time should be visually
indicated somehow.
Original issue reported on code.google.com by [email protected]
on 16 Feb 2011 at 1:51
Error message is "'ascii' codec can't decode byte 0xe5 in position 111: ordinal
not in range(128)"
Original issue reported on code.google.com by [email protected]
on 26 Jul 2011 at 10:17
We can do this easily by doing it async. CheckFeed adds a feed-checking task to
an internal in-memory queue. A sub-thread handles these tasks, and stores
parsed feeds as in memory object structures. Then it sends a ProcessFeed
message to the queue, which does the actual processing. Should be trivial to
implement.
Original issue reported on code.google.com by [email protected]
on 26 Jul 2011 at 6:18
Need to stop repeating the same boilerplate stuff in all templates.
Original issue reported on code.google.com by [email protected]
on 24 Jul 2011 at 10:39
Basically, serve the list with N extra rows. For each vote, send an async
message to the server (basically the same as now), then hide that row, and
unhide another row.
Original issue reported on code.google.com by [email protected]
on 20 Jul 2011 at 6:55
If the recalculation of story points as well as downloading of new stories were
done in a queue-based approach it would make the code a lot more amenable to
deployment in a cloud-type/service-type environment.
Original issue reported on code.google.com by [email protected]
on 7 Feb 2011 at 7:19
The current memory footprint is excessive, mainly because both the original
summary text for a post is stored in memory, and also the text vector.
Original issue reported on code.google.com by [email protected]
on 6 Feb 2011 at 8:33
Or perhaps not? Given that all feeds are public now. Need to consider this more
carefully.
Original issue reported on code.google.com by [email protected]
on 15 Sep 2011 at 11:13
What it says.
Original issue reported on code.google.com by [email protected]
on 29 Jul 2011 at 10:13
We have a feed in the system now that uses https.
Original issue reported on code.google.com by [email protected]
on 25 Jul 2011 at 12:24
Basically, aggregate the scores for posts across all users (simple SQL sum?)
and show a couple of pages of that.
Original issue reported on code.google.com by [email protected]
on 20 Jul 2011 at 6:33
If a vote request causes an error, the error display appears. If the next vote
request succeeds, the error display remains.
Original issue reported on code.google.com by [email protected]
on 20 Aug 2011 at 6:04
This will make it much easier to implement issue 18.
Original issue reported on code.google.com by [email protected]
on 20 Jul 2011 at 6:55
It's worth a shot, because this operation is just way too slow right now.
Original issue reported on code.google.com by [email protected]
on 26 Jul 2011 at 4:07
If the message is already in the queue there's no point in adding it again.
This could prevent some kinds of issues.
Original issue reported on code.google.com by [email protected]
on 28 Jul 2011 at 4:57
what i see now:
25 feed items on one page
what i'd like:
to configure f.ex. one of 10, 15, 20, 25
(motivation: other feature request on marking all on one page as read)
Original issue reported on code.google.com by ingvald
on 8 Aug 2011 at 8:32
Need a statistics screen showing basic information about the amount of data in
the system.
Original issue reported on code.google.com by [email protected]
on 25 Jul 2011 at 12:36
Stories often get a datetime before the present, and so are listed at "60
minutes", then after a good while they bounce down to "n minutes" and increase
normally after that. Presumably this is a timezone issue somewhere.
Original issue reported on code.google.com by [email protected]
on 25 Jul 2011 at 2:07
The UFP would probably be a far better alternative than the RSS library
currently in whazzup.
Original issue reported on code.google.com by [email protected]
on 6 Feb 2011 at 7:41
At the moment whazzup needs to know in advance what the format of the feed it
is loading is. This restriction needs to be removed by auto-detecting the feed
format.
Original issue reported on code.google.com by [email protected]
on 6 Feb 2011 at 7:40
Remove feeds which never load after a given grace period. Should perhaps notify
subscribers by email?
Original issue reported on code.google.com by [email protected]
on 24 Jul 2011 at 1:25
What it says. One user has asked for this.
Original issue reported on code.google.com by [email protected]
on 23 Jul 2011 at 1:12
This date format occurs in one of the feeds and is causing errors, so support
it.
Original issue reported on code.google.com by [email protected]
on 12 Feb 2011 at 12:16
At the moment we can only run a single dbqueue instance, as we might otherwise
get contention over the gdbm files. Need to consider how we can get around this.
Original issue reported on code.google.com by [email protected]
on 24 Jul 2011 at 10:40
Right now we're using an unholy hack to throw web.py off its stride and stop,
causing Python to stop. We don't respect Ctrl-C, either.
Original issue reported on code.google.com by [email protected]
on 21 Feb 2011 at 2:34
At the moment the interaction between whazzup.py and the rest of the system is
a real mix-up of different kinds of calls. Need to build a proper abstraction
that makes this separation cleaner. Ideally it should make it possible to swap
in and out different queueing mechanisms.
Original issue reported on code.google.com by [email protected]
on 12 Feb 2011 at 12:15
The URIs change, which causes the feed identifier to also change, and the code
cannot handle this at the moment. A better solution is to assign IDs
automatically and use those everywhere.
Original issue reported on code.google.com by [email protected]
on 7 Feb 2011 at 7:35
It should be fairly easy to get this to run on Google AppEngine, which would
make it possible to set it up as a free service for anyone to use. Would be
cool to try it out.
Original issue reported on code.google.com by [email protected]
on 12 Feb 2011 at 12:17
what i see now:
a list of 25 feed items, where i can mark one item as read at a time.
what i'd like:
(in addition to fewer/ configurable number of items pr page, )
ability to mark all items on a page as read in one operation.
Original issue reported on code.google.com by ingvald
on 8 Aug 2011 at 8:29
Occasionally DB connections break, and when they do they never get repaired
again. Need to come up with some way to detect and fix.
Original issue reported on code.google.com by [email protected]
on 20 Aug 2011 at 11:20
Sometimes, when reading a feed, the worker thread hangs for several minutes
waiting for the server to time out. This means the entire worker thread hangs
for the same time, blocking all updates. Should redo the code so that it times
out more quickly.
Original issue reported on code.google.com by [email protected]
on 11 Feb 2011 at 6:34
One way would be to dump stats to HTML (for viewing through the web interface)
at intervals, and to keep comparable stats for several such intervals.
Some possible measures: number of tasks in the queue, time to process a task,
wait time for a task in the queue before it's finally processed. Showing max,
min, average, and right now values for these measures might work.
Original issue reported on code.google.com by [email protected]
on 25 Jul 2011 at 7:11
We can handle this by parsing the HTML to find the link to the RSS/Atom feed
and subscribing to that instead.
Original issue reported on code.google.com by [email protected]
on 20 Jul 2011 at 6:31
Don't allow a feed with the same URL as a previous feed to be added to the
database.
Original issue reported on code.google.com by [email protected]
on 19 Feb 2011 at 3:42
What it says.
Original issue reported on code.google.com by [email protected]
on 24 Jul 2011 at 10:41
Turns out SysV IPC message queues suck big time, so we need to change over to
Gearman.
Original issue reported on code.google.com by [email protected]
on 20 Jul 2011 at 1:25
Now that we have OPML we should be good citizens and also support export to OPML
Original issue reported on code.google.com by [email protected]
on 20 Feb 2011 at 6:12
A declarative, efficient, and flexible JavaScript library for building user interfaces.
๐ Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. ๐๐๐
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google โค๏ธ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.