Search Results

Search found 6783 results on 272 pages for 'elastic cache'.

Page 150/272 | < Previous Page | 146 147 148 149 150 151 152 153 154 155 156 157  | Next Page >

  • it started with a broken package

    - by Inteloidl.com
    all started when i tried to updated to the latest lts, but the problen is that the update interumpted and some broken packages apeared, i used synaptics to repair, but it wont repair, iv tried some console from some forum topic ... and in the end when i try to open synaptic or update managre this is what apears any ideas : E: Dynamic MMap ran out of room. Please increase the size of APT::Cache-Limit. Current value: 25165824. (man 5 apt.conf) E: Error occurred while processing mantis (NewVersion1) E: Problem with MergeList /var/lib/apt/lists/c.archive.ubuntu.com_ubuntu_dists_lucid_universe_binary-i386_Packages E: The package lists or status file could not be parsed or opened. E: _cache-open() failed, please report. W: Unable to munmap

    Read the article

  • Specific website not working on my connection [closed]

    - by Tsury
    For some reason, the website http://www.woopra.com/ doesn't work well (shows as plain html, no css/images/JS) on my computer and my Android tablet (they both share the same internet connection). It does however, work on my WP 7.5 phone (3G). When I try to set a wifi hotspot and use the tablet it still doesn't work (tried to clean cache). I tried FF, Chrome and IE (multiple versions). What is going on here? I'm clueless! Thanks.

    Read the article

  • ubuntu 10.10 maverick Zentyal . do anyone know where i can find its ? / download and install its ?

    - by santosamaru
    I want to installing Zentyal , do anyone know about the maverick version ? cause i got the news , that saying webmin is not available anymore and i have problem with its so that is why i try to change the webmin into Zentyal , but the problem is as i can find the version that they have is For Ubuntu 10.04 (Lucid) mine was ubuntu 10.10 maverick. i have problem with speed connections, that is why i need Zentyal so its will help me to build the squid to cache the webpage so its will help to make faster , lol, please your idea bout problem ,this, thanks.

    Read the article

  • 14.04 missing "/etc/init.d/ufw"? my firewall never auto starts

    - by Aquarius Power
    I need to know how to fix the missing "/etc/init.d/ufw" file, is it some package or some command? I used the gufw to enable it, but on reboot my firewall was still off... I created a symlink /etc/init.d/ufw -> /lib/init/upstart-job but I could not make it work like start ufw. I found this file /lib/ufw/ufw-init, it looks like a init.d file! can I copy or symlink it there? Additional (optional) questions: How to find what package has that file? apt-cache search didnt work.. Can we safely create such a script? Any idea why is it missing? Obs.: my /etc/ufw/ufw.conf has ENABLED=yes (but seems useless..)

    Read the article

  • Google I/O 2010 - GWT Linkers target HTML5 WebWorkers & more

    Google I/O 2010 - GWT Linkers target HTML5 WebWorkers & more Google I/O 2010 - GWT Linkers target HTML5 Web Workers, Chrome Extensions, and more GWT 301 Matt Mastracci At its core GWT has a well-defined and customizable mechanism -- called Linkers -- that controls exactly how GWT's compiled JavaScript should be packaged, served, and run. This session will describe how to create linkers and explains some of the linkers we've created, including a linker that turns a GWT module into an HTML5 Web Worker and one that generates an HTML App Cache manifest automatically. For all I/O 2010 sessions, please go to code.google.com From: GoogleDevelopers Views: 6 1 ratings Time: 59:59 More in Science & Technology

    Read the article

  • EC2 Auto-Scaling with Spot and On-Demand Instances?

    - by platforms
    I'm looking to optimize the cost of our auto-scaling EC2 groups by having them launch spot instances instead of on-demand instances. What I really want is to be able to keep some servers in the group as on-demand instances, regardless of what happens to the spot instance pricing market. Then I want any additional servers in the group, above my configured minimum, to be spot instances. I'm generally OK with the delay in adding servers via spot requests. I can't seem to find any way to do this and I've tried to scour the AWS documentation. It appears that an ASG can either be on-demand or spot, but not a hybrid. I could possibly manually add an on-demand instance to the Elastic Load Balancer assigned to the auto-scaling group, but then the load of that server would not be factored into the auto-scaling measurements and triggers. I suppose I could enter a ridiculously high bid price in order to ensure that I always get the servers I need, but then I look at the pricing history and see occasional large spikes. The AWS documentation is at odds with itself, since in one place it says that if you enter a server minimum, that number is "ensured" to be there. But then when you read about spot instances, there are no assurances. The price differential for spot is compelling, so I'd like to leverage that as much as I can while still maintaining an always-on baseline. Is this possible?

    Read the article

  • Dependency problem while missing package is already installed

    - by hakermania
    I am trying to install a program but I am getting a dependency error. The error clearly points out: Dependency is not satisfiable: libc6-amd64 (>= 2.14) I went on to investigate and I found out that I have 2.19 version installed, actually: alex@MaD-pc:~$ apt-cache policy libc6-amd64 libc6-amd64:i386: Installed: 2.19-0ubuntu6 Candidate: 2.19-0ubuntu6 Version table: *** 2.19-0ubuntu6 0 500 http://us.archive.ubuntu.com/ubuntu/ trusty/main i386 Packages 100 /var/lib/dpkg/status Why am I getting this error if I already have this package? I also should probably mention that the system is 100% up to date. I run the updates and upgrades, restarted the system and then tried to install the package again with the same error popping up. Edit 1: I am using amd64 but I have installed some 32-bit libraries required by some program installed via wine if I recall correctly.

    Read the article

  • SQLServerCentral Webinar Series #21 - Forgotten Rings & Other Monitoring Stories

    Most common monitoring metrics are important and useful, especially over time, but they can fall short. How do you gather information to determine, for example, if you have buffer cache pressure? Register now for the free webinar. Wednesday, October 17 2012 4:00pm - 5:00pm BST Are you sure you can restore your backups? Run full restore + DBCC CHECKDB quickly and easily with SQL Backup Pro's new automated verification. Check for corruption and prepare for when disaster strikes. Try it now.

    Read the article

  • Results stored in a session - good idea?

    - by Nick
    To give a bit of background, lets say it's a generic results page, which is paginated so there are X results per page. Generally to do this, I have two queries on the page: to get the total number of results to get the results, limiting by the correct page's resultset However, recently I've been trying to cut down on the queries the site is making, and I thought one way to do this would be to only do the query if any parameters to the page have changed (except of course the page number)? This would then cache all the result id's in a session, which can be sliced when I need to return the correct resultset for that page. I was trying to look around the net to see if there are downsides of this method, but I've found very little information about it. Has anyone done this before? Is it a good idea?

    Read the article

  • Updating query results

    - by Francisco Garcia
    Within a DDD and CQRS context, a query result is displayed as table rows. Whenever new rows are inserted or deleted, their positions must be calculated by comparing the previous query result with the most recent one. This is needed to visualize with an animation new or deleted rows. The model of my view contains an array of the displayed query results. But I need a place to compare its contents against the latest query. Right now I consider my model view part of my application layer, but the comparison of two query result sets seems something that must be done within the domain layer. Which component should cache a query result and which one compare them? Are view models (and their cached contents) supposed to be in the application layer?

    Read the article

  • ec2 instance won't boot!

    - by TheToolBox
    So I had a server lose connection while updating to 14.04LTS. Foolishly, it ended up getting rebooted, and now I'm here. I disconnected the volume and mounted it on another system, chrooted it, and updated the kernel. Still no dice. Any idea what the problem could be? Thanks in advance! The instance log is below: ******************* BLKFRONT for device/vbd/2049 ********** backend at /local/domain/0/backend/vbd/3005/2049 Failed to read /local/domain/0/backend/vbd/3005/2049/feature-barrier. Failed to read /local/domain/0/backend/vbd/3005/2049/feature-flush-cache. 16777216 sectors of 512 bytes ************************** [H[J Booting 'Ubuntu 14.04 LTS, memtest86+' root (hd0) Filesystem type is ext2fs, using whole disk kernel /boot/memtest86+.bin ============= Init TPM Front ================ Tpmfront:Error Unable to read device/vtpm/0/backend-id during tpmfront initialization! error = ENOENT Tpmfront:Info Shutting down tpmfront xc: error: panic: xc_dom_core.c:621: xc_dom_find_loader: no loader found: Invalid kernel xc_dom_parse_image returned -1 close(3) Error 9: Unknown boot failure Press any key to continue...

    Read the article

  • How to run ubuntu-tweak's janitor automatically?

    - by Eliran Malka
    My aim is to have the janitor running at startup, with a pre-configured profile (e.g. clear redundant packages and browser cache). The website is lacking any documentation or usage instructions, and I could not find any information on this here as well. I tried, naturally, to start ubuntu-tweak from the command line, hoping additional API exists that will come through in this (allegedly) simple task. I only got as far as: ubuntu-tweak -f janitor which is a step in the right direction, but what's still missing is a command for the clear action. Is such a command available, or is there any better way of achieving the desired behavior?

    Read the article

  • Is there an apt command to download a deb file to the current directory?

    - by Lekensteyn
    I am often interested in the installation triggers (postinst, postrm) or certain parts (/usr/share or /etc) of packages. Currently, I am running the next command to retrieve the source code: apt-get source [package-name] The downside is, this file is often much bigger than the binary package and does not reflect the installation tree. Right now, I am downloading the packages through http://packages.ubuntu.com/: Search for [package-name] Select the package Click on amd64/i386 for download Download the actual file This takes too long for me and as someone who really likes the shell, I would like to do something like the next (imaginary) command: apt-get get-deb-file [package-name] I could not find something like this in the apt-get manual page. The most close I found was the --download-only switch, but this puts the package in /var/cache/apt/archives (which requires root permissions) and not in the current directory.

    Read the article

  • What is proper way to get specific version of package

    - by Peter
    I am using Ubuntu 14.04 LTS I need to use curl package. But it comes in version 7.35.0 with known SSL bug (already fixed in next version) I either need lower version <= 7.33.0 or higher >= 7.36.0 to make it works, but how can I do it? Do I need to manually install curl? I checked synaptics, "Force Version" feature is greyed out. I also checked apt-cache showpkg curl, there is no other version available. I also tried to add "proposed" repositories, no new curl version either. What should I do? I am thinking about getting .deb package from previous ubuntu release...

    Read the article

  • Can't access my partitions after installing Ubuntu

    - by Manaf Al-Sarraf
    I installed Ubuntu 13.4 replace with win 8 after installation complete I can't access my partitions and receive a message like this : Error mounting /dev/sda5 at /media/manaf/01CD6C0800CEE670: Command-line `mount -t "ntfs" -o "uhelper=udisks2,nodev,nosuid,uid=1000,gid=1000,dmask=0077,fmask=0177" "/dev/sda5" "/media/manaf/01CD6C0800CEE670"' exited with non-zero exit status 14: The disk contains an unclean file system (0, 0). Metadata kept in Windows cache, refused to mount. Failed to mount '/dev/sda5': Operation not permitted The NTFS partition is in an unsafe state. Please resume and shutdown Windows fully (no hibernation or fast restarting), or mount the volume read-only with the 'ro' mount option." Thanks for helping :)

    Read the article

  • Nagios DNX plugins

    - by danneh3826
    I'm toying with the idea of multiple Nagios instances setup to monitor our infrastructure. I've looked at all the various methods of distributed Nagios checks, and I think DNX comes out the closest. DNX handles failure of worker nodes, that's fine. What happens if the main DNX server fails though? Is there a way to replicate the server too? I'm using AWS EC2 primarily, so I can utilise Elastic Load Balancing for the web UI, but I need to be able to handle the AZ where the monitoring server is to fail over, and essentially for a second to pick up the checking load (active/passive, active/active, so long as it doesn't fail completely) The other thing I'm trying to solve is an issue with routing. What I'd like is to have multiple nodes report a fault before Nagios confirms it as critical. Not the NRPE checks, as they're pretty self explanitory, but things more like check_ping. I often have routing issues out of AWS to certain datacenters, so Nagios can often report bad/no ping/timeout as a critical issue, even though the machine in question is working fine. Would it be possible to have a setup where a worker complains a service check is critical, and have a second worker node (positioned in another datacenter/AZ) also report the service as critical before the Nagios central server issues a critical alert? I realise I might be asking a bit much (how far down the line do you go setting up failover systems before it starts to get ridiculous), however surely someone must have thought of this scenario when developing DNX?

    Read the article

  • Why does Chrome crash on websites using the Awesome font?

    - by harry
    Since github using webfront for icons, every time I open github website that tab will crash. But other websites are not crash. I did everything I can, like disable all extensions, clean all cache, even create a new user profile, still not working. I'm using Ubuntu linux 12.04, last stable google chrome browser, always up to date. I searched from google and not found any one has the issue like mine. And I found any website using Font-Awesome my chrome tab will crash if I open that website.

    Read the article

  • Are HTTP requests cached? [closed]

    - by nischayn22
    Many HTTP requests are sent repeatedly by browsers on almost every page load, such as requesting the jQuery .js file etc. Since these are already used on too many sites doesn't modern browsers keep a cache for this? I am thinking of a system where the browser has a cached copy of the .js file used very very frequently. On a new request for the .js file, it sends the server a request for a hash of the .js file (provided the server can reply to that) and compares the returned hash with the cached copy's hash... rest is intuitive.

    Read the article

  • Unable to locate essential development tools Ubuntu 11.04

    - by Anita 7
    I'm using Ubuntu 11.04 (VMware). I aim to implement OpenMP. Im using gcc 4.5 compiler. I tried to install it by using the command sudo apt-get install gcc 4.5. Afterwards I proceed with gcc -fopenmp foo.c BUT the output was: gcc: foo.c: No such file or directory gcc: no input files –. Now I tried to install the package by using : ubuntu@ubuntu:~$ sudo apt-get install essential Reading package lists... Done Building dependency tree Reading state information... Done E: Unable to locate package essential. I also tried apt-cache search essential and after that sudo apt-get install essential-dev But the same error again, E: Unable to locate package essential-dev Any solution,please? Do I need to download any package? What should I do? Thank you in advance :))

    Read the article

  • Keyboad layout: In 13.10, modified symbols do not apply

    - by MPi
    I like to tweak my Colemak layout a bit, so I changed /usr/share/X11/xkb/symbols/us to contain my changes. Sure, they get lost on an upgrade, but that is not very often. After upgrading to 13.10, this does not work anymore. I changed the file, but the changes are not applied. Neither when I use the settings program, nor when I issue setxkbmap 'us(colemak)' directly. Where is this data stored now, is there some kind of cache?

    Read the article

  • How to CURL and avoid timeout death (Twitter Down) [migrated]

    - by David
    Twitter is down right now, and one of my site's home pages relies on getting data from Twitter (relies is the problem - it should be more of an accessory feature, as it just shows follow count from its feed). Here's the code in question: function socials_Twitter_GetFollowerCount($username) { $method = function () use ($username) { return file_get_contents('https://api.twitter.com/1/users/show.json?screen_name='.$username.'&include_entities=true'); }; $json = cache('bmdtwitter', 3600, $method, false); $json = json_decode($json, true); return intval($json['followers_count']); } What is a good way to make it so if Twitter is down (or not responsive for some reasonable amount of time), our site doesn't appear to be down (I think the timeout maybe defaulting to 30-60 seconds or more).

    Read the article

  • Navigation for ASP.NET Web Forms project published on codeplex

    Navigation for ASP.NET Web Forms manages movement and data passing between aspx Pages in a unit testable manner. There is no Client-side logic, so it works in all browsers, and no Server-side cache, so it works with the browser back button.Features include loosely coupled Pages, typed data passing, empty code-behinds, context-sensitive bread crumb trail, ASP.NET Data binding integration, automatic ASP.NET Ajax history navigation and many more.The source code, binaries and comprehensive documentation...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • SQL SERVER Data Pages in Buffer Pool Data Stored in MemoryCache

    This will drop all the clean buffers so we will be able to start again from there. Now, run the following script and check the execution plan of the query. Have you ever wondered what types of data are there in your cache? During SQL Server Trainings, I am usually asked if there is any [...]...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • Visual Studio Load Tests Virtual Users Simulation

    - by Eldar
    Hello, I'm currently working on writing a load testing application that takes advantage of Load Test using Visual Studio 2010. The load test will simulate 20 users on the same machine, and I need some data to be shared in-memory between all simulated users. I was suprised I couldn't find documentation answering the following question: What seperates each virtual user's running context from the other? Does each virtual user runs the tests in its own process? Maybe in its own app domain? Or just on its own thread? I need to know because if each user is running tests in its own process then all the in-memory cache isn't shared and is created for each user instead of one time for all of them, which is bad for me.

    Read the article

  • Set modified date = created date or null on record creation?

    - by User
    I've been following the convention of adding created and modified columns to most of my database tables. I also have been leaving the modified column as null on record creation and only setting a value on actual modification. The other alternative is to set the modified date to be equal to created date on record creation. I've been doing it the former way but I recent ran into one con which is seriously making me think of switching. I needed to set a database cache dependency to find out if any existing data has been changed or new data added. Instead of being able to do the following: SELECT MAX(modified) FROM customer I have to do this: SELECT GREATEST(MAX(created), MAX(modified)) FROM customer The negative being that it's a more complicated query and slower. Another thing is in file systems I believe they usually use the second convention of setting modified date = created date on creation. What are the pros and cons of the different methods? That is, what are the issues to consider?

    Read the article

< Previous Page | 146 147 148 149 150 151 152 153 154 155 156 157  | Next Page >