Friday, June 3, 2011

Re: Caching large datasets on Client or Server?

On Jun 3, 12:55 pm, Gambo <mark.vanv...@gmail.com> wrote:
> Hi there,
>
> I am struggling in finding a good solution for a caching method in my
> application. The client is sometimes retrieving more than 5000 rows
> which will be displayed in a pageable grid for now. In my first
> solution everytime the users clicked on next page the server only send
> the data snapshot of this side which works ok. But it always gets the
> complete dataset from the database.
>
> Now we are adding sorting which should also be done on the server. My
> question is now what is the best strategy to follow here?
>
> I thought about caching the complete 5000 rows on server side but I
> really dont know how to do this as the server is stateless. The only
> thing I have is httpSession to identify but i dont know if its a good
> idea to store this cache in a httpsession.
>
> Another problem is that my app supports a filtering option which could
> change the data of 5000 rows so the cache would be invalid.
>
> Is it a good idea to cache to rows on client side?

It is a question of latency. Wait for every page or wait longer
initially and experience ibstant response thereafter. If each row is a
domain object, you get a lot of extensibility - sorting, row-specific
right-click menu, filtering, categorizing..... etc. You can make the
user forget that he is viewing in a browser.

J.Ganesan
www.DataStoreGwt.com


>I am not sure as
> the sorting would be done in Javascript right?
>
> Thanks for your help

--
You received this message because you are subscribed to the Google Groups "Google Web Toolkit" group.
To post to this group, send email to google-web-toolkit@googlegroups.com.
To unsubscribe from this group, send email to google-web-toolkit+unsubscribe@googlegroups.com.
For more options, visit this group at http://groups.google.com/group/google-web-toolkit?hl=en.

0 Comments:

Post a Comment

Subscribe to Post Comments [Atom]

<< Home


Real Estate