Search Results

Search found 34922 results on 1397 pages for 'form load'.

Page 639/1397 | < Previous Page | 635 636 637 638 639 640 641 642 643 644 645 646  | Next Page >

  • Understanding the 'High Performance' meaning in Extreme Transaction Processing

    - by kyap
    Despite my previous blogs entries on SOA/BPM and Identity Management, the domain where I'm the most passionated is definitely the Extreme Transaction Processing, commonly called XTP.I came across XTP back to 2007 while I was still FMW Product Manager in EMEA. At that time Oracle acquired a company called Tangosol, which owned an unique product called Coherence that we renamed to Oracle Coherence. Beside this innovative renaming of the product, to be honest, I didn't know much about it, except being a "distributed in-memory cache for Extreme Transaction Processing"... not very helpful still.In general when people doesn't fully understand a technology or a concept, they tend to find some shortcuts, either correct or not, to justify their lack-of understanding... and of course I was part of this category of individuals. And the shortcut was "Oracle Coherence Cache helps to improve Performance". Excellent marketing slogan... but not very meaningful still. By chance I was able to get away quickly from that group in July 2007* at Thames Valley Park (UK), after I attended one of the most interesting workshops, in my 10 years career in Oracle, delivered by Brian Oliver. The biggest mistake I made was to assume that performance improvement with Coherence was related to the response time. Which can be considered as legitimus at that time, because after-all caches help to reduce latency on cached data access, hence reduce the response-time. But like all caches, you need to define caching and expiration policies, thinking about the cache-missed strategy, and most of the time you have to re-write partially your application in order to work with the cache. At a result, the expected benefit vanishes... so, not very useful then?The key mistake I made was my perception or obsession on how performance improvement should be driven, but I strongly believe this is still a common problem to most of the developers. In fact we all know the that the performance of a system is generally presented by the Capacity (or Throughput), with the 2 important dimensions Speed (response-time) and Volume (load) :Capacity (TPS) = Volume (T) / Speed (S)To increase the Capacity, we can either reduce the Speed(in terms of response-time), or to increase the Volume. However we tend to only focus on reducing the Speed dimension, perhaps it is more concrete and tangible to measure, and nicer to present to our management because there's a direct impact onto the end-users experience. On the other hand, we assume the Volume can be addressed by the underlying hardware or software stack, so if we need more capacity (scale out), we just add more hardware or software. Unfortunately, the reality proves that IT is never as ideal as we assume...The challenge with Speed improvement approach is that it is generally difficult and costly to make things already fast... faster. And by adding Coherence will not necessarily help either. Even though we manage to do so, the Capacity can not increase forever because... the Speed can be influenced by the Volume. For all system, we always have a performance illustration as follow: In all traditional system, the increase of Volume (Transaction) will also increase the Speed (Response-Time) as some point. The reason is simple: most of the time the Application logics were not designed to scale. As an example, if you have a while-loop in your application, it is natural to conceive that parsing 200 entries will require double execution-time compared to 100 entries. If you need to "Speed-up" the execution, you can only upgrade your hardware (scale-up) with faster CPU and/or network to reduce network latency. It is technically limited and economically inefficient. And this is exactly where XTP and Coherence kick in. The primary objective of XTP is about designing applications which can scale-out for increasing the Volume, by applying coding techniques to keep the execution-time as constant as possible, independently of the number of runtime data being manipulated. It is actually not just about having an application running as fast as possible, but about having a much more predictable system, with constant response-time and linearly scale, so we can easily increase throughput by adding more hardwares in parallel. It is in general combined with the Low Latency Programming model, where we tried to optimize the network usage as much as possible, either from the programmatic angle (less network-hoops to complete a task), and/or from a hardware angle (faster network equipments). In this picture, Oracle Coherence can be considered as software-level XTP enabler, via the Distributed-Cache because it can guarantee: - Constant Data Objects access time, independently from the number of Objects and the Coherence Cluster size - Data Objects Distribution by Affinity for in-memory data grouping - In-place Data Processing for parallel executionTo summarize, Oracle Coherence is indeed useful to improve your application performance, just not in the way we commonly think. It's not about the Speed itself, but about the overall Capacity with Extreme Load while keeping consistant Speed. In the future I will keep adding new blog entries around this topic, with some sample codes experiences sharing that I capture in the last few years. In the meanwhile if you want to know more how Oracle Coherence, I strongly suggest you to start with checking how our worldwide customers are using Oracle Coherence first, then you can start playing with the product through our tutorial.Have Fun !

    Read the article

  • Why does Live-USB boot freeze at the moving dots?

    - by Caleb
    I am currently trying to boot into Ubuntu 12.04 x86 using a USB. I used Untetbootin to do all that boot crap and what not. But for some reason Ubuntu wont load pass that screen with the dots and the purple background. What happens is the dots all light up once, then they go back to white then they all light up at the same time and then it just gets stuck there. Ive even tried using an x64 copy (since my CPU supports it) but have no luck with that either. Ive even tried different DISTRIBUTIONS of Linux and all have failed with loading. Please help

    Read the article

  • Assets.getBytes returns null in test environment

    - by ashes999
    I'm using the latest Haxe (2.10), NME (3.4.3), and MUnit. I've written some unit tests that need to fetch bitmap data from SWF symbols. The first step is to actually load the SWF data. To do this, I use NME's getByteArray along with the swf library, like so: var blah:SWF = new SWF(Assets.getBytes("assets/swf/test.swf")); The call to Assets.getBytes returns null when I'm running this under MUnit. When running my actual game code, I'm able to get the byte array (and consequentially, instantiate the SWF class). Am I doing something wrong? What am I missing? Edit: My directory structure is: . (root .\assets .\assets\*.png (other images) .\assets\swf\*.swf (SWFs) .\Source\*.hx (source code) .\Test\*.hx (tests)

    Read the article

  • How can a collection class instantiate many objects with one database call?

    - by Buttle Butkus
    I have a baseClass where I do not want public setters. I have a load($id) method that will retrieve the data for that object from the db. I have been using static class methods like getBy($property,$values) to return multiple class objects using a single database call. But some people say that static methods are not OOP. So now I'm trying to create a baseClassCollection that can do the same thing. But it can't, because it cannot access protected setters. I don't want everyone to be able to set the object's data. But it seems that it is an all-or-nothing proposition. I cannot give just the collection class access to the setters. I've seen a solution using debug_backtrace() but that seems inelegant. I'm moving toward just making the setters public. Are there any other solutions? Or should I even be looking for other solutions?

    Read the article

  • Oracle NoSQL Database Exceeds 1 Million Mixed YCSB Ops/Sec

    - by Charles Lamb
    We ran a set of YCSB performance tests on Oracle NoSQL Database using SSD cards and Intel Xeon E5-2690 CPUs with the goal of achieving 1M mixed ops/sec on a 95% read / 5% update workload. We used the standard YCSB parameters: 13 byte keys and 1KB data size (1,102 bytes after serialization). The maximum database size was 2 billion records, or approximately 2 TB of data. We sized the shards to ensure that this was not an "in-memory" test (i.e. the data portion of the B-Trees did not fit into memory). All updates were durable and used the "simple majority" replica ack policy, effectively 'committing to the network'. All read operations used the Consistency.NONE_REQUIRED parameter allowing reads to be performed on any replica. In the past we have achieved 100K ops/sec using SSD cards on a single shard cluster (replication factor 3) so for this test we used 10 shards on 15 Storage Nodes with each SN carrying 2 Rep Nodes and each RN assigned to its own SSD card. After correcting a scaling problem in YCSB, we blew past the 1M ops/sec mark with 8 shards and proceeded to hit 1.2M ops/sec with 10 shards.  Hardware Configuration We used 15 servers, each configured with two 335 GB SSD cards. We did not have homogeneous CPUs across all 15 servers available to us so 12 of the 15 were Xeon E5-2690, 2.9 GHz, 2 sockets, 32 threads, 193 GB RAM, and the other 3 were Xeon E5-2680, 2.7 GHz, 2 sockets, 32 threads, 193 GB RAM.  There might have been some upside in having all 15 machines configured with the faster CPU, but since CPU was not the limiting factor we don't believe the improvement would be significant. The client machines were Xeon X5670, 2.93 GHz, 2 sockets, 24 threads, 96 GB RAM. Although the clients had 96 GB of RAM, neither the NoSQL Database or YCSB clients require anywhere near that amount of memory and the test could have just easily been run with much less. Networking was all 10GigE. YCSB Scaling Problem We made three modifications to the YCSB benchmark. The first was to allow the test to accommodate more than 2 billion records (effectively int's vs long's). To keep the key size constant, we changed the code to use base 32 for the user ids. The second change involved to the way we run the YCSB client in order to make the test itself horizontally scalable.The basic problem has to do with the way the YCSB test creates its Zipfian distribution of keys which is intended to model "real" loads by generating clusters of key collisions. Unfortunately, the percentage of collisions on the most contentious keys remains the same even as the number of keys in the database increases. As we scale up the load, the number of collisions on those keys increases as well, eventually exceeding the capacity of the single server used for a given key.This is not a workload that is realistic or amenable to horizontal scaling. YCSB does provide alternate key distribution algorithms so this is not a shortcoming of YCSB in general. We decided that a better model would be for the key collisions to be limited to a given YCSB client process. That way, as additional YCSB client processes (i.e. additional load) are added, they each maintain the same number of collisions they encounter themselves, but do not increase the number of collisions on a single key in the entire store. We added client processes proportionally to the number of records in the database (and therefore the number of shards). This change to the use of YCSB better models a use case where new groups of users are likely to access either just their own entries, or entries within their own subgroups, rather than all users showing the same interest in a single global collection of keys. If an application finds every user having the same likelihood of wanting to modify a single global key, that application has no real hope of getting horizontal scaling. Finally, we used read/modify/write (also known as "Compare And Set") style updates during the mixed phase. This uses versioned operations to make sure that no updates are lost. This mode of operation provides better application behavior than the way we have typically run YCSB in the past, and is only practical at scale because we eliminated the shared key collision hotspots.It is also a more realistic testing scenario. To reiterate, all updates used a simple majority replica ack policy making them durable. Scalability Results In the table below, the "KVS Size" column is the number of records with the number of shards and the replication factor. Hence, the first row indicates 400m total records in the NoSQL Database (KV Store), 2 shards, and a replication factor of 3. The "Clients" column indicates the number of YCSB client processes. "Threads" is the number of threads per process with the total number of threads. Hence, 90 threads per YCSB process for a total of 360 threads. The client processes were distributed across 10 client machines. Shards KVS Size Clients Mixed (records) Threads OverallThroughput(ops/sec) Read Latencyav/95%/99%(ms) Write Latencyav/95%/99%(ms) 2 400m(2x3) 4 90(360) 302,152 0.76/1/3 3.08/8/35 4 800m(4x3) 8 90(720) 558,569 0.79/1/4 3.82/16/45 8 1600m(8x3) 16 90(1440) 1,028,868 0.85/2/5 4.29/21/51 10 2000m(10x3) 20 90(1800) 1,244,550 0.88/2/6 4.47/23/53

    Read the article

  • Connected to wifi, but quits loading pages after a random amount of time

    - by Collin
    I'm using Ubuntu 12.10, on a TOSHIBA Satellite L755 laptop, and using a Cradlepoint MBR-1000 wireless router. The wifi supposedly works fine on my sisters' and mom's laptops, and my dad uses a wired connection through the router. Nobody else has the problem. It shows that I'm connected and everything. The network manager GUI shows that everything is going just dandy... but it's not. I'll be able to load a page or two, then I have to disconnect, and reconnect to the router to get it to work again for a couple minutes. I noticed when I was downstairs, close to the router, this problem seemed to go away (at least, I was connected for a couple of hours without running into this problem). When I run ping -n 8.8.8.8 it shows this while the connection is still running fine 64 bytes from 8.8.8.8: icmp_req=1187 ttl=47 time=665 ms and whe

    Read the article

  • Canonical links for huge websites

    - by Florin
    Let's say I have 5 products that are identical but the product code, the product color specifications and the product image. The title, meta and description are identical (by the way the color is in a select form). I made 4 products link canonical to the 1 that is the master based on many factors. If the master becomes inactive or without a stock one product from the other 4 will become the new master and the rest will become canonical to it. The question is if that by becomeing master from canonical will the site suffer a penalty from Google or it will work just fine? What will Google think about this strategy?

    Read the article

  • Best way for a mailing list in Android

    - by Asgard
    I want to create a mailing list function based on an address book included in the application, (NOT the user's contacts) for example I have the addresses [email protected] [email protected] [email protected] [email protected] [email protected] every email addresses, must be associated with a specific section, the various section should be selected through a checkbox list. every section may have more than an email. I have yet created a working email sender that support the simultaneous sending to many email but I don't know what is the best way to store and load the address in the application and how realize this.

    Read the article

  • How to retrieve data from a corruped volume

    - by explorex
    Hi, My Ubuntu 10.10 just crashed(probably due to hardware error and in the end I was getting error like Unknown filesystem ..... grub> .. GRUB console before i could take some action) and i reinstalled the same version form USB stick. I had ubuntu installed in ext4 file system and I am also having the same filesystem in the same hard disk on different drive. When I try to access my previous filesystem, i get error Error mounting: mount: wrong fs type, bad option, bad superblock on /dev/sda6, missing codepage or helper program, or other error In some cases useful info is found in syslog - try dmesg | tail or so I had some important files in the previous volume, I don't know how to retrieve them. And what are the chances that I would get the same outcome (hardware error)? Please help me!

    Read the article

  • Complete RESTful API debugging/testing tool

    - by vartec
    I'm looking for the most complete tool, preferably portable GUI or browser plugin to test RESTful API. What I need is: GET/POST/DELETE/PUT support multiple file uploads as fields (multipart/form-data) file uploads as body Extra points for: possibility to save multiple configurations and use them to pre-fill parameters OAuth support nice JSON response formatting Currently I'm using 3 tools: Chrome REST Console extension — My favorite, very nicely done. Has OAuth. However the functionality missing for me is sending file as a body of the request; Cannot send multiple files; Firefox Poster add-on — Quite nice, but the functionality it's missing for file as POST fields parameters; Also cannot send multiple files; cURL — can do anything, but it's quite tedious to use it from command line.

    Read the article

  • Smartassembly 5: it lives! Early Access builds now available

    - by Bart Read
    I'm pleased to announce that, late last week, we put out the first early access build for Smartassembly 5, Red Gate's fantastic code protection and error reporting tool, which we acquired last September. You can download it via: http://www.red-gate.com/messageboard/viewforum.php?f=116 It's obviously pretty early days, so please do not try to use this to protect a production application, but we've already done a lot of work in some key areas: We're simplifying and streamlining the licensing model (you won't see this yet, but a lot of the work on this has already been done). We've improved usability of the product, with a better menu, reordering of project settings, and better defaults. We've also fixed a load of bugs, which I'll let Alex blog about in more detail. On a slightly more trivial level, the curly braces are also no more. Over the coming weeks, we'll be adding more improvements, and starting usability tests. If you're interested in getting involved in the latter, please drop an email to [email protected].

    Read the article

  • Ubuntu 12.04 desktop crash [closed]

    - by David Mannock
    12.04 is stable under a light load, but not under intensive use of Gaussian 09 vB1. Looks like a heating issue, but psensor says that all is well on 32-cores @56C. Similar results for 64 cores. Machine shuts down after 2-3h. Syslog shows shut down. Whoopsie crash reporter sends in report. After 259 updates on the weekend, I am left wondering what the heck is wrong with this release? My answer would be "EVERYTHING!". Can someone help me do some systematic checks on this OS and hardware.

    Read the article

  • Moving large website to new CMS - URL changes

    - by herrherr
    Hi, I was wondering if you have any tipps on the following situation. I'm going to move a large website to a new Content Management System, here are some details on the site: online news magazine with roughly 3,000 articles domain age: 10 years online in the current form since May 2010 indexed pages: ~10.000 percent of search engine traffic: under 10% Unfortunately a custom-tailored CMS was used for the site. The performance, reliability and SEO capabilites have been really bad, so we are moving to a new and proven open source CMS. All the articles will be kept as they are, but the URL structure as well as the structure of the HTML templates will be changed. What I wanted to do now is to actually create 301 redirects for all articles from the old to the new schema, i.e: Old: www.example.com/en/html/news/detail/title-of-the-article/ New: www.example.com/category/title-of-article.html Is this a proven way to do something like this? If not, can you recommend a way that has worked for you? Thanks :)

    Read the article

  • Computer refuses to boot either ubuntu disk I'm using

    - by xtraorange
    I'm attempting to set up an Amahi server, and thus have taken an older PC and wiped it clean using DBAN. I'm now in the process of installing Ubuntu, but for some reason the computer refuses to recognize either form of Ubuntu I throw at it, even though it recognizes the Windows XP CD fine. Things it can't be: Boot order - I've tripple checked A bad DVD - I've booted the live CD off of my Mac just fine. A drive issue - the drive boots other cd's just fine, I've tested both an XP and the ultimate boot disk for windows... they are booting fine. A drive conflict - I have unplugged the only hard drive, leaving the DVD drive alone, still no recognition. I've tried both Ubuntu 12 and an older Ubuntu 10 CD, it refuses to recognize either one. It's like my computer has decided it will have nothing to do with Ubuntu. Any help would be greatly appreciated.

    Read the article

  • how to bind data binding source to group radio buttons

    Hi I am working on windows c# application (visual studio 20008).In my form I have 5 Radio buttons ,grouped by group box. I have to bind this radio button to data binding source. this.rdoptPlus.DataBindings.Add(new Binding("Checked", MemberDefinitionsBindingSource, "UnaryOperator", true)); this.rdoptMinus.DataBindings.Add(new Binding("Checked", MemberDefinitionsBindingSource, "UnaryOperator", true)); i am unable to bind single datacolum to 5 radio buttons,please help me  read moreBy masuri suhasiniDid you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • sudo: source: command not found

    - by HorusKol
    I've been updating some of the default profile for bash, and saw from the tutorials I was following that I could reload the new profile with the new environment settings by using: source /etc/bash.bashrc The only thing is - the new environment variables were only available to my current user - and were ignored when I used sudo. They only became available to sudo when I closed my terminal session and rejoined. When I try to use: sudo source /etc/bash.bashrc I get the error: sudo: source: command not found Is there a simple way to load in the new bash profile settings for sudo without having to close the terminal and restart?

    Read the article

  • Error: no such partition help!

    - by John
    Ok so i had Ubuntu installed on dual boot with windows 7, i decided i liked it, btu i did not want the dual boot option so i decided to uninstall it. no whenever i turn on my computer the "error: no such partition" screen shows up and i cant do anything at all. i tried restarting it and booting from a windows recovery disc to fix the mbr but my bios no wont load i cant change the boot order i cant manually boot or anything, it just goes straight to the "error: no such partition" screen. can someone please help me?

    Read the article

  • multi-clients web application,should I use custom user controls or a common user control

    - by ValidfroM
    Say my company is going to build a complicated asp.net web form education system. One of the module is web based registration. To make it flexiable, we decide to use user control(ascx) with rule-engine (work flow) regulating all business logic behide them. Thus in future,for different clients, we can simply config basic existing rules or adding new rules.(Rules stored in db or XML per client). Now the question is how to deal with the user controls (ascx)? My opinion is for different client build diffrent user control from scratch. other voice is like reuse existing user controls.

    Read the article

  • Why does sound stop working after a while?

    - by badp
    I don't know how to reproduce this problem, because I don't regularly play music or sound. All I know is that, sometimes, I'll load a video (from youtube or from a local file) and there will be no sound. Everything looks fine software wise: Rebooting always fixes. aplay, paplay and pals give no error message I'm not in the audio group, as advised The device exists and appears in use: $ lsof /dev/snd/by-path/pci-0000\:00\:1b.0 COMMAND PID USER FD TYPE DEVICE SIZE/OFF NODE NAME pulseaudi 17313 badp 23u CHR 116,10 0t0 7628 /dev/snd/by-path/../controlC0 pulseaudi 17313 badp 30u CHR 116,10 0t0 7628 /dev/snd/by-path/../controlC0 Restarting pulseaudio or alsa seems to do no good. What is wrong here?

    Read the article

  • Unity Desktop not loading and ccsm method not working

    - by slimy-spork
    I installed Ubuntu 13.04 x86_64 on my HP Pavillion DV7-6c60us. It installs fine and works until I update. I run the software updater and reboot as it requests after. When my computer comes back on I get the desktop background but the rest of the Unity desktop doesn't load. I've tried the ccsm method and re-enabling but it just de-enables itself. I've also tried installing the gnome desktop but that does nothing for me either. I really want to switch to Ubuntu but this is causing issues. P.S. I've also tried using 12.10 and 12.04 with no dice.

    Read the article

  • Shopping cart for service providers?

    - by uos??
    From my limited exposure, it seems to me that most shopping cart/eCommerce platforms are specifically for products-based retailers. On several occasions now, I've been asked about ecommerce solutions for service providers. That is, it's basically just a single product with payment but no shipping, and highly configurable "product". Any recommendations for a cost-efficient solution (high feature coverage) for such a web platform? Requirements: .NET No/suppressed product catalog A service customization selection form Payment (probably PayPal with accountless credit card processing) Guest purchases (no site account required) Email confirmation Customer service -facing control panel It's hard to search for such a product because I get "web service based ecommerce software" and so on clouding up the results.

    Read the article

  • Is there a cross-platform special directory I can use for game save files?

    - by Suds
    I'm developing with LWJGL and Java on a Windows 7 laptop. I've successfully set up saving to the %appdata%\gamename\saves\ or long form c:\users\user\appdata\roaming\gamename\saves\ folder by using File dir = new File(System.getenv("APPDATA") + "\\gamename\\saves\\");. I have hobbyist level experience with Linux, and zero experience with OSX. My game will be fully cross platform. Is System.getenv("APPDATA"); cross platform? If so, where does it point to on Linux or OSX? Is there a best practices alternative that I should use?

    Read the article

  • How can I link to callback functions in Lua such that the callbacks will be updated when the scripts are reloaded?

    - by Raptormeat
    I'm implementing Lua scripting in my game using LuaBind, and one of the things I'm not clear on is the logistics of reloading the scripts live ingame. Currently, using the LuaBind C++ class luabind::object, I save references to Lua callbacks directly in the classes that use them. Then I can use luabind::call_function using that object in order to call the Lua code from the C++ code. I haven't tested this yet, but my assumption is that if I reload the scripts, then all the functions will be redefined, BUT the references to the OLD functions will still exist in the form of the luabind::object held by the C++ code. I would like to be able to swap out the old for the new without manually having to manage this for every script hook in the game. How best to change this so the process works? My first thought is to not save a reference to the function directly, but maybe save the function name instead, and grab the function by name every time we want to call it. I'm looking for better ideas!

    Read the article

  • Own an iPad, iPhone, or Touch? Add GWB to your Home

    - by Staff of Geeks
    I know it is goofy to geek out about such things, but I am a pretty goofy geek.  Today we published support for the Apple icons for devices like the iPhone, iPad, or iPod Touch.  Now when you add Geekswithblogs.net to your home screen (via the little + in Safari) it will load the icon to the left instead of a screenshot of the site.  We have been using my iPhone and iPad to surf to GWB for awhile and we were tired of that screenshot.  Simple Google/Bing search and a little Photoshop CS5 (Amazing!!!) work and there it is.  Next step we are taking in the Apple iPad/Phone arena is getting our developer kit and seeing what possiblities there are for Geekswithblogs.net and a custom App.  Got any ideas? Here is a shot of my iPad 3G: Also, would anyone be interested in a Geekswithblogs.net custom background for the iPhone or iPad?  I can put something together real quick if there are any takers.

    Read the article

  • Windows 7 not booting from Ubuntu 12.10 boot menu

    - by Raj Inevitable
    Am having windows 7 in one hard drive. I installed Ubuntu in second hard drive. Am configured BIOS to boot second hard drive first (Ubuntu OS). and then i updated grub, so its shows windows 7 in the boot list. I can boot in to ubuntu, but i can't boot into windows 7, its shows error A disk read error occured Press Ctrl+Alt+Del to restart . then am configured BIOS again to load windows 7 first, it shows windows 7 and ubuntu in the boot list, windows 7 is working, but i can't boot into ubuntu.. Help to solve this problem.. i want dual boot from any one of the drive...

    Read the article

< Previous Page | 635 636 637 638 639 640 641 642 643 644 645 646  | Next Page >