Search Results

Search found 6949 results on 278 pages for 'loading'.

Page 147/278 | < Previous Page | 143 144 145 146 147 148 149 150 151 152 153 154  | Next Page >

  • What is a good motherboard for the Intel i7 series processor?

    - by jasondavis
    I am lost on choosing a good miotherboard at a decent price for my Intel i7-930 CPU. I have read bad things about all of them so far. Issues with RAM not working correctly, Windows not loading with USB keyboards plugged in and all kinds of nightmare stories. Here is my needs. - 6g/b SATA 3 - USB 3.0 - Cheaper is better - Support for up to 24gb of RAM, I will start out with only 12gb though - I will use 64bit v. of Windows 7 so the 4gb RAM limit should not be a problem from my OS - Should be able to boot even if I have a USB keyboard plugged in. If you have experience with a motherboard that meets these specs, please do tell about it. I appreciate any help, I am stuck right now on picking a good board.

    Read the article

  • Tomcat "connection interrupted" with ssl

    - by Mike Thomsen
    I can access Tomcat on port 8080, but not on 8443. When I try o get there, this is the error I get in Firefox: The connection was interrupted The connection to the.fqdn.com:8443 was interrupted while the page was loading. This is my connector: <Connector port="8443" protocol="org.apache.coyote.http11.Http11NioProtocol" SSLEnabled="true" maxThreads="150" scheme="https" secure="true" clientAuth="want" sslProtocol="TLS" keystoreFile="C:\temp\keystore.jks" keystorePass="changeit"/> I have the CA key in the jre's cacerts file. The server cert was generated using EJBCA and should be signed properly. Any suggestions on what is going on?

    Read the article

  • Trying to install SawMill and getting the following error:

    - by Itai Ganot
    [root@sawmill sawmill]# ./sawmill ./sawmill: error while loading shared libraries: libldap-2.3.so.0: cannot open shared object file: No such file or directory Using yum provides libldap_r-2.3.so.0 i found that the package which includes this file is: compat-openldap-2.3.43-2.el6.i686 . After installing it i still get the error. If i use locate, i can find the file in /usr/lib, so I tried to create a symbolic link to the file from /usr/lib to /usr/lib64 but i still get the same error. I also tried setting LD_LIBRARY_PATH=/usr/lib/ and LD_LIBRARY_PATH=/usr/lib64 but it doesn't allow me to run the sawmill installation script. Anyone knows how to solve this issue?

    Read the article

  • How do I setup a local DNS server on Mac OS Lion?

    - by Peter Kovacs
    I had some serious lag to resolve website address and sometimes things simply wouldn't load (pages kept loading for 5+ minutes without even a timeout error). So I had setup a local dns server/cache using BIND on Leopard and Snow Leopard. Now that I have Lion, i have the same problem, but the instructions no longer apply to Lion and I can't find a way to do it. Has anyone attempted to do this? Are there viable alternatives for DNS servers on OS X 10.7? For those who are wondering I already tried several external DNS server. Only my computer has this issue on the network.

    Read the article

  • How to update a keyboard layout DLL on Windows?

    - by user87690
    Information about keyboard layout is stored in keyboard layout DLLs on Windows. One can make a custom layout using a tool like Microsoft Keyboard Layout Creator (http://msdn.microsoft.com/en-us/goglobal/bb964665.aspx). It creates the dynamical library and also an installer which places the DLL at a right place and also puts necessary information to registry so the layout can be used. Now my question. How can one update his custom layout? One could uninstall the layout and install the new version which effectively replaces the DLL file. However this alone doesn't work because it seems that the old copy of the DLL stays loaded somewhere and is used when one sets the layout instead of loading new version. So is there a way how to tell the system that its “keyboard layout cache” is invalid? Of course I could reboot the whole system but I'd like to avoid it as it's annoying and seems to be poor design to reboot whole system just to reload a DLL.

    Read the article

  • Minimal Fedora Installation

    - by MA1
    I am working on a Fedora 13 minimal iso image with no Desktop environment. I just need wxPython support. I removed the gnome from kickstart and build the iso image using livecd-creator. But now my application is not loading as gnome is gone. Earlier my application(myapp.desktop) was placed in /usr/share/gnome/autostart/ and starts automatically. So what should i do now the run my application? To run my wxPython appalication should i have to install display manager(xdm, gdm, kdm etc..)? If i install gdm, it takes lots of space. what should i do? In short i need Fedora 13 minimal iso image with no desktop environment and with wxPython support.

    Read the article

  • Why is Firefox so slow and heavy?

    - by Tony
    For some reason, when I go to links the pages seem slow and heavy. It also has a lot of lag spikes between page loads. Basically it seems to freeze then load it all at once fast. I'm currently using Firefox 25. But when I use the same Chrome version, it seems to be very fast and smooth page loading. The CPU it takes on average is about 400,000k. Extensions: iMacros Leethax Ad Block Plus 2.4 Ad Block Plus Pop-up Addon 0.9.1 Computer stats: 6 GB RAM Windows 7 Acer Aspire Laptop 500 GB HDD Intel Core i4-2370M How do I make Firefox load like Google Chrome, without much freezing?

    Read the article

  • Dropped Dell XPS has hard-drive trouble

    - by Alex B.
    Yesterday, my mom dropped her laptop to the floor and got the blue screen of death after trying to boot the system. I was able to start a Fedora live CD and get some of her stuff off the hard drive, but I cannot seem to be able to install Windows on it. The installation starts loading files and then the computer turns off. I am thinking that she might need a new hard drive. Any ideas? Edit: I actually was able to boot the Windows XP installation but it is saying that no hard disk is detected. How can this be possible if I was able to mount the drive on fedora yesterday?

    Read the article

  • Why is the size of windows off by 226x238 if defined via the Window Rules?

    - by Bobby
    I have installed Sawfish 1.8.2 from source on my new Ubuntu 12.04 installation following the Debian instructions, but I had this problem also with the stock 1.5.3. Whenever I define dimensions in the Window Rules for a window, the size is off by exactly 226x238 pixels, which means that 100x100 turns into 326x328. That's a very odd behavior, given that Sawfish is saving and loading the dimensions of the windows correctly (if saved via the window menu). Some additional system information: $ uname -a Linux Dagon 3.2.0-24-generic-pae #39-Ubuntu SMP Mon May 21 18:54:21 UTC 2012 i686 i686 i386 GNU/Linux $ sawfish --version sawfish version 1.8.2 nvidia proprietary driver, 9600GT Two monitors, 1920x1080 + 1440x900 in one session. Positionng the windows is working fine, only the dimensions are off by that odd number. Does somebody have an idea why?

    Read the article

  • nginx static file buffer

    - by Philip
    I have a nfs which several frontend-servers are connected to for making the files stored on the nfs available for http downloads. It looks like I have problems with the way apache is serving the files, there seems to be a very small buffer or no buffer at all which results in a lot disk seeks. I did some testing with loading the whole requested file into memory at once and serve it to the client from memory. With this technique I need less disk seeks for a download stream. Since I don't want to implement this by myself for production use I thought that I could maybe use nginx for that because the documentation says that it uses buffers for static file serving. Is it possible to increase the buffer size to a few mb, if so which config parameter do I have to change for this? Has anyone experience with large buffers for static file serving? Is there a better way to reduce disk seeks?

    Read the article

  • How do I set the umask for files and directories created from the GUI in MacOS X Lion (10.7)?

    - by Avry
    I've set my umask in my .bashrc file to 007. Any files created on the command line after loading my bashrc file respects this setting. I want to be able to set the umask to 007 for any files created using non-command line apps. This document talks about setting the umask via launchd. And it kind of works. If I follow these directions I can change the default permissions on a GUI created file from rw-r--r-- to rw-rw---- but the directories still are not group writeable (i.e. I want them to be rwxrwx--- but they are rwxr-x--- instead) The analog on Linux would be /etc/login.defs as the place to set the umask. What do I change in order for the umask to be set properly (i.e. the way I want it)?

    Read the article

  • Apache on Windows random long wait times

    - by Jaxbot
    I have a development machine with Apache installed as a service on Windows. The installation is fresh out of the box, with no changes to configuration aside from adding the PHP module. From day one, I've had a problem that looks like this: Essentially, Apache is freezing for about 11 seconds before replying on random requests. This appears to happen more frequently when the host hasn't been connected to in a while, but this is not always the case. I've eliminated MySQL, PHP, and the specific application; this long wait problem will occur even when loading a static file such as favicon.ico. Thus, the only factor remaining is Apache, which is freezing for consistently around 10-11 seconds before replying. The problem is not the DNS problem that many people point to; as you can see, the DNS lookup is instant, and the problem occurs both on localhost and 127.0.0.1. Thanks for the time.

    Read the article

  • Generating cache file for Twitter rss feed

    - by Kerri
    I'm working on a site with a simple php-generated twitter box with user timeline tweets pulled from the user_timeline rss feed, and cached to a local file to cut down on loads, and as backup for when twitter goes down. I based the caching on this: http://snipplr.com/view/8156/twitter-cache/. It all seemed to be working well yesterday, but today I discovered the cache file was blank. Deleting it then loading again generated a fresh file. The code I'm using is below. I've edited it to try to get it to work with what I was already using to display the feed and probably messed something crucial up. The changes I made are the following (and I strongly believe that any of these could be the cause): - Revised the time difference code (the linked example seemed to use a custom function that wasn't included in the code) Removed the "serialize" function from the "fwrites". This is purely because I couldn't figure out how to unserialize once I loaded it in the display code. I truthfully don't understand the role that serialize plays or how it works, so I'm sure I should have kept it in. If that's the case I just need to understand where/how to deserialize in the second part of the code so that it can be parsed. Removed the $rss variable in favor of just loading up the cache file in my original tweet display code. So, here are the relevant parts of the code I used: <?php $feedURL = "http://twitter.com/statuses/user_timeline/#######.rss"; // START CACHING $cache_file = dirname(__FILE__).'/cache/twitter_cache.rss'; // Start with the cache if(file_exists($cache_file)){ $mtime = (strtotime("now") - filemtime($cache_file)); if($mtime > 600) { $cache_rss = file_get_contents('http://twitter.com/statuses/user_timeline/75168146.rss'); $cache_static = fopen($cache_file, 'wb'); fwrite($cache_static, $cache_rss); fclose($cache_static); } echo "<!-- twitter cache generated ".date('Y-m-d h:i:s', filemtime($cache_file))." -->"; } else { $cache_rss = file_get_contents('http://twitter.com/statuses/user_timeline/#######.rss'); $cache_static = fopen($cache_file, 'wb'); fwrite($cache_static, $cache_rss); fclose($cache_static); } //END CACHING //START DISPLAY $doc = new DOMDocument(); $doc->load($cache_file); $arrFeeds = array(); foreach ($doc->getElementsByTagName('item') as $node) { $itemRSS = array ( 'title' => $node->getElementsByTagName('title')->item(0)->nodeValue, 'date' => $node->getElementsByTagName('pubDate')->item(0)->nodeValue ); array_push($arrFeeds, $itemRSS); } // the rest of the formatting and display code.... } ?> ETA 6/17 Nobody can help…? I'm thinking it has something to do with writing a blank cache file over a good one when twitter is down, because otherwise I imagine that this should be happening every ten minutes when the cache file is overwritten again, but it doesn't happen that frequently. I made the following change to the part where it checks how old the file is to overwrite it: $cache_rss = file_get_contents('http://twitter.com/statuses/user_timeline/75168146.rss'); if($mtime > 600 && $cache_rss != ''){ $cache_static = fopen($cache_file, 'wb'); fwrite($cache_static, $cache_rss); fclose($cache_static); } …so now, it will only write the file if it's over ten minutes old and there's actual content retrieved from the rss page. Do you think this will work?

    Read the article

  • Elastix, how to MOVE files from one server to other server?

    - by yudayyy
    In my office, i have to schedule for moving a file from one computer to other computer (Both are using Elastix). My idea is using cron, scp, and rm to do this. So here are the script that i use: scp -r /home/data/* [email protected]:/home/data1 && rm -r /home/data/* That script did the copy, but not remove the source file. I already read this question: Hov to _MOVE_ files with scp? The problem is, the computer doesn't have an internet connection. So i cannot install rsync on my elastix computer. yum install rsync Loaded plugins: fastestmirror Loading mirror speeds from cached hostfile and then it freezes. Any idea how to do this?

    Read the article

  • Installing Java 1.5 on Ubuntu?

    - by StackedCrooked
    I already have Java 1.6, but I need to test something with 1.5. I have downloaded the .bin file from http://java.sun.com/javase/downloads/index_jdk5.jsp using the Sun Download Manager. Now I want to create a deb file from this bin file: $ fakeroot make-jpkg java_ee_sdk-5_01-linux.bin Creating temporary directory: /tmp/make-jpkg.Zpm1Y7LbZ0 Loading plugins: blackdown-j2re.sh blackdown-j2sdk.sh common.sh ibm-j2re.sh ibm-j2sdk.sh j2re.sh j2sdk-doc.sh j2sdk.sh j2se.sh sun-j2re.sh sun-j2sdk-doc.sh sun-j2sdk.sh Detected Debian build architecture: i386 Detected Debian GNU type: i486-linux-gnu No matching plugin was found. Removing temporary directory: done How can I fix the "No matching plugin was found." error? Update I downloaded jdk-1_5_0_22-linux-amd64.bin from the archive page and ran Linux installer. It works fine.

    Read the article

  • Rails development environment Resque.enqueue does not create jobs

    - by anton evangelatov
    I am having the same problem like Rails custom environment Resque.enqueue does not create jobs , but the solution there doesn't work for me. I'm using Resque for a couple of asynchronous jobs. It works just fine for the staging environment, but for some reason it stopped working on development environment. For example, if I run the following: $ rails c development > Resque.enqueue(MyLovelyJob, 1) Nothing is enqueued. I check Resque using resque-web If I run it on staging - it works just fine. $ rails c staging > Resque.enqueue(MyLovelyJob, 1) I have tried to duplicate the 2 environment, and they seem to use absolutely the same configurations (database.yml , config/environment , etc.), but development is still not working. If I do > Resque.enqueue(UpdateInstancesData, 2) > => true > Resque.info > => { > :pending => 0, > :processed => 0, > :queues => 0, > :workers => 1, > :working => 0, > :failed => 0, > :servers => [ > [0] "redis://127.0.0.1:6379/0" > ], > :environment => "development" > } Any suggestions where to look in order to debug this? I am running the application via foreman. My Procfile looks like: faye: rackup faye.ru -s thin -E production worker1: bundle exec rake resque:work QUEUE=* VERBOSE=1 worker2: bundle exec rake resque:work QUEUE=* VERBOSE=1 clock: bundle exec rake resque:scheduler VERBOSE=1 web: bundle exec rails s For staging, as mentioned, everything works and the log from foreman is: 17:03:42 clock.1 | 2013-06-26 17:03:42 Reloading Schedule 17:03:42 clock.1 | 2013-06-26 17:03:42 Loading Schedule 17:03:42 clock.1 | 2013-06-26 17:03:42 Scheduling logging_test 17:03:42 clock.1 | 2013-06-26 17:03:42 Schedules Loaded 17:03:43 worker2.1 | *** Starting worker ttttt-mbp.local:69573:* 17:03:43 worker2.1 | *** Registered signals 17:03:43 worker2.1 | *** Running before_first_fork hooks 17:03:43 worker1.1 | *** Starting worker ttttt-mbp.local:69572:* 17:03:43 worker1.1 | *** Registered signals 17:03:43 worker2.1 | *** Checking another_queue 17:03:43 worker2.1 | *** Checking anotherqueue 17:03:43 worker2.1 | *** Checking statused 17:03:43 worker2.1 | *** Found job on statused 17:03:43 worker2.1 | *** got: (Job{statused} | LoggingTest | ["57e89a1c1b24ce6866bcf5d0e1c07f01", {}]) 17:06:30 clock.1 | 2013-06-26 17:06:30 queueing LoggingTest (logging_test) 17:06:33 worker1.1 | *** Checking another_queue 17:06:33 worker2.1 | *** Checking another_queue 17:06:33 worker1.1 | *** Checking anotherqueue 17:06:33 worker2.1 | *** Checking anotherqueue 17:06:33 worker1.1 | *** Found job on anotherqueue 17:06:33 worker1.1 | *** got: (Job{anotherqueue} | LoggingTest | ["0d976869a945766e0cfeca83e7349305", {}]) 17:06:33 worker1.1 | *** resque-1.24.1: Processing anotherqueue since 1372259193 [LoggingTest] 17:06:33 worker1.1 | *** Running before_fork hooks with [(Job{anotherqueue} | LoggingTest | ["0d976869a945766e0cfeca83e7349305", {}])] 17:06:33 worker1.1 | *** resque-1.24.1: Forked 69955 at 1372259193 17:06:33 worker2.1 | *** resque-1.24.1: Forked 69956 at 1372259193 17:06:33 worker1.1 | *** Running after_fork hooks with [(Job{anotherqueue} | LoggingTest | ["0d976869a945766e0cfeca83e7349305", {}])] 17:06:33 worker1.1 | JOB :: LoggingTest 17:06:33 worker1.1 | 55555 17:06:33 worker1.1 | *** done: (Job{anotherqueue} | LoggingTest | ["0d976869a945766e0cfeca83e7349305", {}]) whereas for development it doesn't seem to enqueue and then find the job. If there is a job already in the queue (pending, left over from staging environment) the workers from development don't process it. 17:01:23 clock.1 | 2013-06-26 17:01:23 Reloading Schedule 17:01:23 clock.1 | 2013-06-26 17:01:23 Loading Schedule 17:01:23 clock.1 | 2013-06-26 17:01:23 Scheduling logging_test 17:01:23 clock.1 | 2013-06-26 17:01:23 Scheduling update_instances_data 17:01:23 clock.1 | 2013-06-26 17:01:23 Schedules Loaded 17:03:10 clock.1 | 2013-06-26 17:03:10 queueing LoggingTest (logging_test) 17:03:14 worker1.1 | *** Checking another_queue 17:03:14 worker2.1 | *** Checking another_queue 17:03:14 worker1.1 | *** Checking anotherqueue 17:03:14 worker2.1 | *** Checking anotherqueue 17:03:14 worker1.1 | *** Checking statused 17:03:14 worker2.1 | *** Checking statused

    Read the article

  • ESXi 4 failed to find boot device

    - by phil
    I had exsi 3.5 running just fine, the upgrade to 4 failed ( i thought it would so everything is backed up ). but it did boot just could not import the data store So i did a fresh install of ESXi 4. Now it says failed to find boot device. Not sure why. i have wiped the disk, the controller is a adaptec 3405, it worked fine in 3.5 and when i upgrade to 4 it did boot. Now it just hangs at loading vfat. then the failed to find boot device. Any ideas?

    Read the article

  • modprobe ndiswrapper - not found

    - by David
    I'm trying to use ndiswrapper on a Slackware 12 (I think) box, but I'm running into a problem with modprobe. Everything I find online says that it should be working, but for some unknown reason it isn't. Here's what I've done so far: Installed ndiswrapper (latest tarball, make, make install) Ran ndiswrapper -i on the WinXP driver for my USB wireless card Ran ndiswrapper -l which tells me the driver is present and the device is present (lsusb also confirms the device is present) Ran ndiswrapper -m which put an alias for wlan0 in /etc/modprobe.d/ndiswrapper.conf Ran depmod -a Ran modprobe ndiswrapper which tells me "FATAL: module ndiswrapper not found" Ran modprobe -l which shows no listing for ndiswrapper I even tossed in a reboot or two while trying various combinations of the above, still nothing. So naturally ifconfig wlan0 up isn't working because the device isn't being created, presumably because the module isn't loading the driver. Does anybody have any suggestions? Everything points to the notion that this should work fine, but modprobe just isn't able to find what it needs. Have I missed an important step?

    Read the article

  • Force Firefox 14 to free memory when opening/closing lots of popups

    - by aknghiem
    I'm currently trying to run some tests on a web application using Selenium IDE with Firefox 14. The tests mainly consist in loading a page containing thousands of links and clicking on every of those links. Of course, each time a popup shows, I tell Selenium to close it and proceed with the remaining links. However, it seems that even if I close the popups, Firefox is not freeing memory. Usually, I end up with Firefox crashing after opening 1500 popups (around 2.5Gb of memory usage). Is there any way to force the browser to free memory? Maybe something I should set in about:config? Or is there a flaw with Selenium? Thanks.

    Read the article

  • Using resize to getScript if above x pixels (jQuery)

    - by user1065573
    So I have an issue with this script. It basically is loading the script on any resize event. I have it working if lets say... - User has window size above 768 (getScipt() and content by .load() in that script) - Script will load (and content) - User for some reason window size goes below 768 (css hides #div) - User re-sizes again above 768, And does not load the script again! (css displays the div) Great. Works right. Now.. (for some cray reason) - User has window size below 768 (Does NOT getScript() or nothing) - User re-sizes above 768 and stops at any px (getScript and content) - User re-sizes again anything above 768 (It getScript AGAIN) I need it to only get script once. Ive used some code from here jQuery: How to detect window width on the fly? (edit - i found this, which he has the same issue of loading a script once.) Problem with function within $(window).resize - using jQuery And others i lost the links to :/ When i first found this problem I was using something like this, then i added the allcheckWidth() to solve problem. But it does not. var $window = $(window); function checkWidth() { var windowsize = $window.width(); ////// LETS GET SOME THINGS IF IS DESKTOP var $desktop_load = 0; if (windowsize >= 768) { if (!$desktop_load) { // GET DESKTOP SCRIPT TO KEEP THINGS QUICK $.getScript("js/desktop.js", function() { $desktop_load = 1; }); } else { } } ////// LETS GET SOME THINGS IF IS MOBILE if (windowsize < 768) { } } checkWidth(); function allcheckWidth () { var windowsize = $window.width(); //// IF WINDOW SIZE LOADS < 768 THEN CHANGES >= 768 LOAD ONCE var $desktop_load = 0; if (!$desktop_load) { if (windowsize < 768) { if ( $(window).width() >= 768) { $.getScript("js/desktop.js", function() { $desktop_load = 1; }); } } } } $(window).resize(allcheckWidth); Now im using something like this which makes more sense? $(function() { $(window).resize(function() { var $desktop_load = 0; //Dekstop if (window.innerWidth >= 768) { if (!$desktop_load) { // GET DESKTOP SCRIPT TO KEEP THINGS QUICK $.getScript("js/desktop.js", function() { $desktop_load + 1; }); } else { } } if (window.innerWidth < 768) { if (window.innerWidth >= 768) { if (!$desktop_load) { // GET DESKTOP SCRIPT TO KEEP THINGS QUICK $.getScript("js/desktop.js", function() { $desktop_load + 1; }); } else { } } } }) .resize(); // trigger resize event }) Ty for future response. Edit - To give an example, in desktop.js i have a gif that loads before "abc" content, that gets inserted by .load(). On resize this gif will show up and the .load() in desktop.js will fire again. So if getScript was only being called once, it should not be doing anything again in desktop.js. Confusing?

    Read the article

  • Why do I get a GRUB error after deleting the Linux partition?

    - by nateify
    I had a dual-boot setup with Windows 7 and Windows Vista. I used Easus Partition Manager and shrank my Windows 7 partition. In the empty space, I installed Linux Mint. I decided I didn't want it anymore, I simply deleted the Linux Mint partition. Now when I boot, all I get is this: GRUB loading. error: no such partition grub rescue> I don't have my Windows 7 install disk because it was pre-installed.

    Read the article

  • Mac OS X Time Machine Restore - Failure?

    - by Rabe
    I've got an late 2009 macbook pro 13' and yesterday I replaced the native hd (new hd: Hitachi; same form factor). I choosed "Restore from Time Machine" in the options menu on the snow leopard install discs and waited several hours to complete. After an reboot the mac shows up a white screen with apple logo and tiny loading animation. Nothing happens. After another restart; the same. Now I cant boot from CD using the C-Key and I have no idea how to fix that problem.

    Read the article

  • Creating a file server - How can I use a large VHD file in Hyper-V? (700GB)

    - by barfoon
    Hey everyone, After a few discussions (here, here, and here), I am still unable to create a simple VM that will be used as a fileserver hosted on my Hyper-V box. I have created a fixed 700GB SCSI drive (.vhd file), as I have learned an IDE drive of this size is not possible. Not to sound too cynical, but its blown me away at how much trouble its been to create a large amount of space and start using it. What is the best way to create a fileserver with a drive of this size hosted on Hyper-V Server 2008, and how can I get it going??? Inclusion of OS, driver, integration tools etc, anything you feel is required would be greatly appreciated. Extra information I am using the stand-alone version of Hyper-V server, and not Windows Server 2008. I have tried loading the Linux Integration Tools (linked in the comments of the last link above) onto a SUSE 11 VM and the installation fails, the machine cannot see the vhd at all. Thanks very much,

    Read the article

  • Centos 6, local yum repo, and multiple versions of the same rpm

    - by Tom Skelley
    I'm trying to set up a really simple local repo. I want to have a basic repo with two versions of only one rpm, so I did: mkdir /packages/x64 copy two rpms to /packages/x64 [root@repo x64]# createrepo --verbose /packages/x64 1/2 - jre-6u37-linux-amd64.rpm 2/2 - jre-7u9-linux-x64.rpm Saving Primary metadata Saving file lists metadata Saving other metadata Added the repo to /etc/yum.repos.d/local.repo But when I do: [root@repo x64]# yum list jre I get: Loaded plugins: fastestmirror, security Loading mirror speeds from cached hostfile Available Packages jre.x86_64 1.7.0_09-fcs local ie it only shows the latest version. I know that they're both in the repo because I've run this: [root@repo x64]# rpm -qp jre-6u37-linux-amd64.rpm jre-1.6.0_37-fcs.x86_64 [root@repo x64]# rpm -qp jre-7u9-linux-x64.rpm jre-1.7.0_09-fcs.x86_64 and when I remove the latter version, and run createrepo again, the former shows up. Most puzzling, what am I missing?

    Read the article

  • Accessing server by dedicated IP address

    - by Sherwin Flight
    I'm having an issue with my hosting provider after migrating to a new account. It's taking some time to get the problem sorted out, so I am hoping someone here can shed some light on the situation. The server is running WHM/cPanel, and the site I am trying to access has a dedicated IP address. When I connect to the server like this http://IP.HERE instead of showing my the website the way I would expect, it is showing the contents of a subfolder. So, while I would expect it to load public_html/ it is loading public_html/somefolder/ instead. Any idea why this is happening instead of showing the sites homepage the way I would expect? EDIT It is not redirecting, so the url is just http://IP.ADDRESS/, but the files listed are from a subfolder. So, it LOOKS at though I went to http://IP.ADDRESS/subfolder, when the URL says it should be showing the main folder contents. When I access the site using the domain name, it works properly, so I assume the document root is set correctly.

    Read the article

< Previous Page | 143 144 145 146 147 148 149 150 151 152 153 154  | Next Page >