Search Results

Search found 18766 results on 751 pages for 'me again'.

Page 274/751 | < Previous Page | 270 271 272 273 274 275 276 277 278 279 280 281  | Next Page >

  • How to find location of installed library

    - by Raven
    Background: I'm trying to build my program but first I need to set up libraries in netbeans. My project is using GLU and therefore I installed libglu-dev. I didn't note location where the libraries were located and now I can't find them.. I've switched to Linux just a few days ago and so far I'm very content with it, however I couldn't google this one out and becoming frustrated.. Is there way to find out where files of package were installed without running installation again? I mean if I got library xxx and installed it some time ago, is there somecommand xxx that will print this info? I've already tried locate, find and whereis commands but either I'm missing something or I just can't do it correctly.. for libglu, locate returns: /usr/share/bug/libglu1-mesa /usr/share/bug/libglu1-mesa/control /usr/share/bug/libglu1-mesa/script /usr/share/doc/libglu1-mesa /usr/share/doc/libglu1-mesa/changelog.Debian.gz /usr/share/doc/libglu1-mesa/copyright /usr/share/lintian/overrides/libglu1-mesa /var/lib/dpkg/info/libglu1-mesa:i386.list /var/lib/dpkg/info/libglu1-mesa:i386.md5sums /var/lib/dpkg/info/libglu1-mesa:i386.postinst /var/lib/dpkg/info/libglu1-mesa:i386.postrm /var/lib/dpkg/info/libglu1-mesa:i386.shlibs Other two commands fail to find anything. Now locate did it's job but I'm sure none of those paths is where the library actually resides (at least everything I was linking so far was in /usr/lib or usr/local/lib). The libglu was introduced just as example, I'm looking for general solution for this problem.

    Read the article

  • Why does StackExchange store images in imgur rather than its own servers? [migrated]

    - by martin's
    I am trying to understand the technical (and business) logic behind taking such an approach. Certainly SE isn't short of server or bandwidth resources. I don't think imgur is a CDN, so that can't be the reason. On the one hand one is giving up local control (meaning your files, your hardware) of the content. On the other, you don't have to use your own bandwidth, storage and resources. Then again, you depend on someone else for the reliability and up-time of your service.

    Read the article

  • .NET access to the GPU for compute purposes

    - by Daniel Moth
    In the distant past I talked about GPGPU and Microsoft's then approach of DirectCompute. Since then of course we now have C++ AMP coming out with Visual Studio 11, so there is a mainstream easier way for developers to access the GPU for compute purposes, using C++. The question occasionally arises of how can a .NET developer access the GPU for compute purposes from their C# (or VB) code. The answer is by interoping from the managed code to a native DLL and in the native DLL use C++ AMP. As a long term .NET developer myself, I can tell you this is straightforward. Sure, there could have been a managed wrapper for C++ AMP, but honestly that is the reason we have interop – it doesn't make much sense to invest resources to solve a problem that is already solved (most developer customers would prefer investments in other areas of Visual Studio!). Besides, interoping from C# to C++ is much easier than interoping to some of the other older approaches of GPGPU programming ;-) To help you get started with the interop approach, Igor Ostrovsky has previously shared the "Hello World" version of interoping from C# to C++ AMP in his blog post: How to use C++ AMP from C# …we then were asked specifically about how to interop from C# to C++ AMP in a Metro style application on Windows 8, so Igor delivered again with this post: How to use C++ AMP from C# using WinRT Have fun! Comments about this post by Daniel Moth welcome at the original blog.

    Read the article

  • What's the strategy to implement a "knowledge base" in my company. [closed]

    - by Oscar Reyes
    In my current work we think we can get benefit from having a knowledge base, so the next time someone has a question/problem etc, that base can be consulted and an answer will show up. Also, it will reduce the risk from having people leaving the company with the knowledge and we would have to start all over again. My question is, what strategy can we follow to implement/buy/get/build/etc this knowledge base? Are there software ready for this? Would it be better to have something build by ourselves ( we have some programmers ) This is an small company ( < 30 ) and the base should be accessible from outside the office ( when the employees are with the customer etc.) so I guess a webapp is in order.

    Read the article

  • USB drives not recognized all of a sudden

    - by Siddharth
    I have tried most of the advice on askubuntu and other sites, usb_storage enable to fdisk -l. But I am unable to find steps to get it working again. lsusb results Bus 001 Device 001: ID 1d6b:0002 Linux Foundation 2.0 root hub Bus 002 Device 001: ID 1d6b:0001 Linux Foundation 1.1 root hub Bus 003 Device 001: ID 1d6b:0001 Linux Foundation 1.1 root hub Bus 004 Device 001: ID 1d6b:0001 Linux Foundation 1.1 root hub Bus 005 Device 001: ID 1d6b:0001 Linux Foundation 1.1 root hub Bus 004 Device 002: ID 413c:3012 Dell Computer Corp. Optical Wheel Mouse Bus 005 Device 002: ID 413c:2105 Dell Computer Corp. Model L100 Keyboard Bus 001 Device 005: ID 8564:1000 dmseg | tail reports [ 69.567948] usb 1-4: USB disconnect, device number 4 [ 74.084041] usb 1-6: new high-speed USB device number 5 using ehci_hcd [ 74.240484] Initializing USB Mass Storage driver... [ 74.256033] scsi5 : usb-storage 1-6:1.0 [ 74.256145] usbcore: registered new interface driver usb-storage [ 74.256147] USB Mass Storage support registered. [ 74.257290] usbcore: deregistering interface driver usb-storage fdisk -l reports Device Boot Start End Blocks Id System /dev/sda1 * 2048 972656639 486327296 83 Linux /dev/sda2 972658686 976771071 2056193 5 Extended /dev/sda5 972658688 976771071 2056192 82 Linux swap / Solaris I think I need steps to install and get usb_storage module working. Edit : I tried sudo modprobe -v usb-storage reports sudo modprobe -v usb-storage insmod /lib/modules/3.2.0-48-generic-pae/kernel/drivers/usb/storage/usb-storage.ko Still no usb driver mounted. Nor does a device show up in /dev. Any step by step process to debug and fix this will be really helpful. Thanks.

    Read the article

  • Microsoft Terminology: .NET C++ vs. traditional C++

    - by Mike Clark
    I've recently been working with a team that's using both .NET C++ and pre-.NET C++. I fully understand the technical differences between the two technologies. However, I sometimes feel like I'm floundering when it comes to the terminology used to differentiate the two. Example: Say we have two projects: ProjectA contains "C++" code that builds a .NET assembly DLL. ProjectB contains Visual C++ code that builds a traditional native Windows DLL. What is the best way to succinctly and terminologically draw a distinction between the two projects? Again, I'm not asking for an in-depth technical description of the differences between the two technologies. I'm just looking for names and labels. This is how, today, I might try to make the distinction when talking to someone: "ProjectA is a managed .NET C++ project" and "ProjectB is an unmanaged native C++ DLL project." However I am not at all certain that this terminology is ideal, or even correct. Please describe what you feel the ideal language to use in this situation (or similar situations) might be. Feel free to motivate your answer.

    Read the article

  • Jump and run HTML5 Game Framework

    - by user1818924
    We're developing a jump and run game with HTML5 and JavaScript and have to build an own game framework for this. Here we have some difficulties and would like to ask you for some advice: we have a "Stage" object, which represents the root of our game and is a global div-wrapper. The stage can contain multiple "Scenes", which are also div-elements. We would implement a Scene for the playing task, for pause, etc. and switch between them. Each scene can therefore contain multiple "Layers", representing a canvas. These Layer contain "ObjectEntities", which represent images or other shapes like rectangles, etc. Each Objectentity has its own temporaryCanvas, to be able to draw images for one entity, whereas another contains a rectangle. We set an activeScene in our Stage, so when the game is played, just the active scene is drawn. Calling activeScene.draw(), calls all sublayers to draw, which draw their entities (calling drawImage(entity.canvas)). But is this some kind of good practive? Having multiple canvas to draw? Each gameloop every layer-context is cleared and drawn again. E.g. we just have a still Background-Layer, … wouldn't it be more useful to draw this once and not to clear it everytime and redraw it? Or should we use a global canvas for example in the Stage and just use this canvas to draw? But we thought this would be to expensive... Other question: Do you have any advice how we could dive into implementing an own framework? Most stuff we find online relies on existing frameworks or they just implement their game without building a framework.

    Read the article

  • Transmission dependencies are not in repository

    - by chocobai
    I'm running Ubuntu 12.04 using Gnome3 through their ppa in launchpad. Went good until now. Today Ubuntu wanted to make some updates and I thought okay. But it didn't want to install all of them. Turned out there were not all dependencies in the ppa. Before I noticed that, I (noob :D) uninstalled (apt-get purge) Transmission in order to reinstall it. Bad Idea, because now I can't install it, even using synaptics. The problem: It seems like the only ppas that have transmission in it are the one from gnome3 team (not all packages included according to synaptics, v2.71) and the official (transmission-common missing, v2.61). So when I try to install a package it says that transmission-common is too new. I can't even find the version that should be somewhere in the official standard Ubuntu ppa. Installing from the Software center doesn't work either. Why aren't all dependencies in the ppas or why can't I see them? How can I install Transmission (stable, newest version) again? And why has the gnome3 team a newer version as the stable ppa of transmission? Are they including beta software? If you need any additional information just tell me. Sorry I think it's quiet confusing what I did there. Thanks in advance. EDIT1 Okay sorry, I should have explained better. When I try to install it tells me something of dependency problems. When I try sudo apt-get install ... it says: Die folgenden Pakete haben unerfüllte Abhängigkeiten: transmission-cli : Hängt ab von: transmission-common (= 2.61-0ubuntu0.12.04.1) aber 2.71-0ubuntu1~precise1 soll installiert werden transmission-daemon : Hängt ab von: transmission-common (= 2.61-0ubuntu0.12.04.1) aber 2.71-0ubuntu1~precise1 soll installiert werden transmission-gtk : Hängt ab von: transmission-common (= 2.61-0ubuntu0.12.04.1) aber 2.71-0ubuntu1~precise1 soll installiert werden E: Probleme können nicht korrigiert werden, Sie haben zurückgehaltene defekte Pakete. Unfortunately I'm using Ubuntu in German.

    Read the article

  • How to prevent laptop screen brightness from changing when un/plugging battery power

    - by Nomad
    When I am using my laptop, I continually adjust the screen's brightness based on the lighting conditions in the room (e.g. how much light is coming in from windows, etc.). But if I unplug the laptop or plug it back in, Windows looks at the default brightness setting in the power profile for "on battery" or "plugged in" and changes the brightness accordingly. This is a jarring experience and then I have to hunt down the ideal brightness for my current situation again, rather than getting on with my work. I would like to make it so that plugging or unplugging the battery is not a trigger that adjusts the screen brightness at all. The screen brightness should only change when I adjust it myself. Does anybody know how this might be accomplished?

    Read the article

  • What are "Missing thread recordng" erros when running fsck -fy?

    - by ohho
    There is some error reported when I run Disk Utility and verify the root volume on my OS X MacBook. So I boot and CMD-S into the shell mode and run /sbin/fsck -fy. Errors are like: ** Checking catalog file. Missing thread record (id = ...) In correct number of thread records ** Checking catalog hierarchy. Invalid volume file count (It should be ... instead of ...) ** Repairing Volume Missing directory record (id = ...) I'd like to know what is the cause of the above errors? Hopefully I will be more careful in the future to prevent them from happening again. p.s. I am using a SSD and so I assume mechanical hard disk error is less likely. Thanks!

    Read the article

  • Dialog box keeps asking for password

    - by hossam-khalili
    I am running Windows 7, Outlook 2007 (Office 2007 pro) and I'm connecting to our Exchange Server 2007 which is part of Small Business Server 2008. Outlook 2007 on one client keeps asking for the password to the remote access URL. If I simply click cancel it's OK for a few minutes. Entering the password and clicking the save PW box does no good. Sometimes clicking cancel results in another dilog box asking the same thing and I may have to click cancel several times to get it to go away for a while. Occasionally Outlook may actually go into a mode where it says it needs the password typed so I click the link which brings the dialog back but simply clicking cancel will make Outlook connect again. Can anyone help me? Thanks!

    Read the article

  • Panning with the OpenGL Camera / View Matrix

    - by Pris
    I'm gonna try this again I've been trying to setup a simple camera class with OpenGL but I'm completely lost and I've made zero progress creating anything useful. I'm using modern OpenGL and the glm library for matrix math. To get the most basic thing I can think of down, I'd like to pan an arbitrarily positioned camera around. That means move it along its own Up and Side axes. Here's a picture of a randomly positioned camera looking at an object: It should be clear what the Up (Green) and Side (Red) vectors on the camera are. Even though the picture shows otherwise, assume that the Model matrix is just the identity matrix. Here's what I do to try and get it to work: Step 1: Create my View/Camera matrix (going to refer to it as the View matrix from now on) using glm::lookAt(). Step 2: Capture mouse X and Y positions. Step 3: Create a translation matrix mapping changes in the X mouse position to the camera's Side vector, and mapping changes in the Y mouse position to the camera's Up vector. I get the Side vector from the first column of the View matrix. I get the Up vector from the second column of the View matrix. Step 4: Apply the translation: viewMatrix = glm::translate(viewMatrix,translationVector); But this doesn't work. I see that the mouse movement is mapped to some kind of perpendicular axes, but they're definitely not moving as you'd expect with respect to the camera. Could someone please explain what I'm doing wrong and point me in the right direction with this camera stuff?

    Read the article

  • Mac Book Refuses to Boot

    - by pas09
    I have a Macbook that has unfortunately died on me. I randomly got a pop up message that said I needed to restart my computer, and once I did, I was greeted with the blinking folder question mark start up screen. I've tried everything, including running Disk Utility repair and reformatting the hard drive and reinstalling OS X. All of my data is backed up and safe, I just need my computer to start again. Before I go off and buy a new hard drive, I wanted to see if there is anything I might have forgotten.

    Read the article

  • How to join video files from terminal?

    - by Leon Vitanos
    I have tried avidemux2_cli, mencoder, ffmpeg, cat.. But this doesn't always work (With the most of the times the error is that the audio codec is not the same) Maybe i put wrong options in the commands. So the commands: cat Sample.avi rrr.avi > complete.avi ffmpeg -i Sample.avi -i output.avi -vcodec copy -acodec copy complete.avi mencoder -ovc lavc -oac copy Sample.avi rrr.avi -o complete.avi avidemux2_cli --audio-codec copy --video-codec copy --output-format avi --load Sample.avi -append output.avi --save video.avi The cat problem is that it doesn't show error but it doesn't work always..Like the complete.avi will be exactly the same with Sample.avi Fmmpeg does nothing. The complete.avi is always the same with Sample.avi Mencoder error: All files must have identical audio codec and format for -oac copy. So the complete.avi is the same with Sample.avi avidemux2_cli there is no error but the complete.avi is again the same with Sample.avi.. So to sum up, all complete.avi are the same with Sample.avi.. And the problem is that they don't have the same audio codec ( i quess ).. Any ideas?

    Read the article

  • Sql Server database logs "missing"

    - by linkerro
    Here's the problem we're having: In the applications that are using the databases we get errors like: The log for database 'redactedDatabaseName' is not available. Check the event log for related error messages. Resolve any errors and restart the database. Cannot close event log because there are still event handle users active. The server log says this: Error: 9001, Severity: 21, State: 1. Context: The server in question is holding databases for a multi-tenant application and lately it's been under quite heavy loads. It has something like 3k small databases on it and these errors have been coming in the last week. We have a work-around right now, which is taking the database offline then bringing it online again. Any ideas why this is happening and how we can fix it?

    Read the article

  • VSS error 12293 after system disk clone (Win2003)

    - by carlpett
    Hi! After cloning a windows 2003 installation from a single drive onto two mirrored drives using Acronis Disk Director, VSS no longer works, filing events 12293 and 7001 when trying to use backup tools, and additionally giving error 0x8004230f when accessing the Shadow copy tab of disk properties. I've google-researched this quite throughly, and found a suggested fix[1]: replacing the MBR signature of the disk. This would cause windows to invalidate old shadow copy information, which supposedly would make it all work again. However, I am a bit nervous over this... Is there a possiblity of messing this up somehow, because of the mbr originating from a single disk install, and now residing on a raid mirror? Has anyone here had this problem and solved it? This method or another? [1] http://kb.backupassist.com/articles.php?aid=2971 (under header Resolution 2)

    Read the article

  • Keep context-configuration when redeploying via Cargo

    - by Björn Pollex
    I am using Tomcat 7 to host a web-application that requires a JNDI datasource to be set up. Because this resource is specific to this application, I would like to configure it inside the application-specific context-descriptor in $CATALINA_BASE/conf/[enginename]/[hostname]/. I am also using Cargo from Maven to deploy the web-application to Tomcat. The problem is that when I do a redeploy with Cargo, it first undeploys the application, before deploying it again. When undeploying it, Tomcat deletes the context-descriptor of the application, so it won't work after redeploying. I could of course package the context-descriptor with the application, but I would like to keep any such container-specifics out of the .war. Another alternative is to configure the datasource in the global context-descriptor, but that too seems wrong, because the datasource is supposed to be exclusive to my application. Is my approach fundamentally wrong? What is the best practice here? Is there any way to prevent Tomcat from deleting the descriptor when undeploying?

    Read the article

  • June SOA Partner Community Newsletter 2011

    - by JuergenKress
    Dear SOA Partner Community, It is summer! Thanks for your great support in FY11; we closed our fiscal year on May 31st 2011. Thanks to you, SOA Specialization is the No.1 middleware Specialization! Here is my list to continue to be successful in FY12: Top 10 to become an successful Oracle SOA & BPM; Partner in FY12 SOA Partner Community member SOA Certified Implementation Specialist &BPM Certified Implementation Specialist SOA Blogs& SOA Books SOA Bootcamp& SOA Bootcamp material BPM Bootcamp& BPM Bootcamp material SOA sales kit& BPM sales kit SOA marketing campaign& BPM marketing kit& submit your reference SOA Solutions catalog profile with your service and success SOA Specialization & BPM Specialization & Specialization eBook SOA & BPM @ OOW Summer time means, summer camps– which we do host this week in Lisbon, you can find the latest impression at twitter, If you missed them. Please visit the OPN Competency Center and the enablement blog to get your local training schedule for a SOA & BPM & WebLogic bootcamp near you! Also in this newsletter you will get to read about the new SOA Governance book, JDevelope and ADF 11gR2, Exalogic Specialization and a wrap up of the SOA & Cloud Symposium. Oracle SOA Partner Adoption EMEA – To read the newsletter please visit http://tinyurl.com/soanewsjune2011 To become a member of the SOA Partner Community please register at http://www.oracle.com/goto/emea/soa Till we meet again! Jürgen KressOracle SOA Partner Adoption EMEA Blog Twitter LinkedIn Mix Forum Wiki Website

    Read the article

  • Oracle Database 12c is available for download now!

    - by Mike Dietrich
    Good things come to those who wait ... finally ... Oracle Database 12c (Oracle 12.1.0.1) is available for download from the Oracle Software Cloud (formerly know as eDelivery) and OTN (Oracle Tech Network) for Linux 64bit (Solaris will follow within the next few hours): eDelivery:Oracle Database 12c (12.1.0.1) for Linux 64bitOracle Database 12c (12.1.0.1) for Solaris SPARC64Oracle Database 12c (12.1.0.1) for Solaris x86. OTN:Oracle Database 12c (12.1.0.1) for Linux 64bitOracle Database 12c (12.1.0.1) for Solaris SPARC64Oracle Database 12c (12.1.0.1) for Solaris x86  . And yes, it will be supported on Oracle Exadata and SuperCluster as well . . And with the release of Oracle Database 12c we are offering you also our NEWUpgrade, Migrate and Consolidate to Oracle Database 12cslide deck with (sorry, we've did it again!) over 500 slides covering: The brand new Parallel Upgrade including new Pre/Post-Upgrade-Fix-Ups The new Full Transportable Export/Import Feature Obviously Oracle Multitenant, which got talked about a lot as Pluggable Databases or Container Databases before Plenty of new parameters, cool and very helpful features and much more ... Download the slides Upgrade, Migrate and Consolidate to Oracle Database 12c And of course, the slide deck will see some updates in the near future -Mike . .

    Read the article

  • [Zend Debugger] Cannot receive start command

    - by tharkun
    Hi I have a WAMP install (Apache 2.2.3, PHP 5.2.8) with a working Zend Optimizer. As soon as I add the php.ini lines for Zend Debugger: [Zend Debugger] zend_extension_manager.debug_server_ts="C:\zenddebugger" zend_debugger.allow_hosts=127.0.0.1/32, 192.168.0.0/16, 192.168.1.0/255 zend_debugger.expose_remotely=always and restart apache, I get a blank page and the following entry in the apache error log. [Zend Debugger] Cannot receive start command [Zend Debugger] Cannot send message Now the incredible thing is that I had it working this morning and now, after reinstalling WAMP there is no way I can get it work again. php.ini is exactly the same as it was before. I would be extremely glad to get help, I'm trying to get a working server environment going for over a month, can you believe that? I tried XAMPP, Zend Server and now WAMP and I'm going crazy slowly but surely.

    Read the article

  • SQL Azure Security: DoS

    - by Herve Roggero
    Since I decided to understand in more depth how SQL Azure works I started to dig into its performance characteristics. So I decided to write an application that allows me to put SQL Azure to the test and compare results with a local SQL Server database. One of the options I added is the ability to issue the same command on multiple threads to get certain performance metrics. That's when I stumbled on an interesting security feature of SQL Azure: its Denial of Service (DoS) detection engine. What this security feature does is that it performs a check on the number of connections being established, and if the rate of connection is too high, SQL Azure blocks all communication from that machine. I am still trying to learn more about this specific feature, but it appears that going to the SQL Azure portal and testing the connection from the portal "resets" the feature and you are allowed to connect again... until you reach the login threashold. In the specific test I was performing, all the logins were successful. I haven't tried to login with an invalid account or password... that will be for next time. On my Linked In group (SQL Server and SQL Azure Security: http://www.linkedin.com/groups?gid=2569994&trk=hb_side_g) Chip Andrews (www.sqlsecurity.com) pointed out that this feature in itself could present an internal threat. In theory, a rogue application could be issuing many login requests from a NATed network, which could potentially prevent any production system from connecting to SQL Azure within the same network. My initial response was that this could indeed be the case. However, while the TCP protocol contains the latest NATed IP address of a machine (which masks the origin of the machine making the SQL request), the TDS protocol itself contains the IP Address of the machine making the initial request; so technically there would be a way for SQL Azure to block only the internal IP address making the rogue requests.  So this warrants further investigation... stay tuned...

    Read the article

  • How to make other computers connected to my workgroup able to browse my web server via my private local domain www.mydomain.local?

    - by Motivated Student
    Background I have a broadband connection. The fiber optic cable running from outside to my building is connected to a converter unit. The output of the converter is connected to a router. The router also provides 8 LAN ports to which the computers are connected. Shortly speaking, all computers are interconnected in a workgroup as opposed to a domain. My computer hosts an IIS web server using a private or local domain name www.mydomain.local. I has added 127.0.0.1 www.mydomain.local to the hosts file in the server C:\Windows\System32\drivers\etc so I can browse the site from within the server. So far so good. How to make other computers connected to my workgroup able to browse my web server via my private local domain www.mydomain.local? Note: Again, I am a newbie. I have no idea what I should do next.

    Read the article

  • Tracking changes to firewall configs?

    - by jmreicha
    Myself and one other indivdual will be taking over some of the daily firewall management duties soon and I'm looking for a way to track changes on our firewall configurations for auditing purposes and need some ideas on a good way to track changes the changes that are made. I don't have a lot of specific criteria but here are some of the basic things I would like to be able to do: Access to previous revisions of firewall configs Access to changes made and by whom When specific changes were made I'm wondering if some sort of revision control software would work here as a way to track the the changes? Or if some other approach would work better for managing the change control in this situation. I'm open to any and all suggestions at this point. EDIT: We are using a Checkpoint pair, one passive one active configuration. I will update again with specific model numbers when I get a chance.

    Read the article

  • All nework interfaces hang for seconds while one interface goes up/down

    - by user3698377
    I am building a client/server application that uses several network interfaces in parallel for redundancy, and I have noticed that while one network interface goes down or goes up, the communication on other interfaces hangs for several seconds. I could reproduce this behavior without my application in a simple way: there are 2 interfaces available on computer 1 ( Ethernet and WiFi ) ping from computer 2 the IP address of the Ethernet connection of computer 1 disconnect the WiFi of computer 1 ping hangs for seconds, and then the packets are traveling again between the 2 computers. The hanging happens as well if I turn back on the WiFi connection on computer 1. It happens as well if I ping the WiFi IP, and turn off/on the Ethernet connection ( or unplug/plug the cable). I am using Linux Ubuntu 12.04 on both computers. Any ideas why is this happening, and if / how can it be avoided?

    Read the article

  • VMWare Player guest does not re-aquire IP on bridged interface after host loses and re-acquires network connectivity

    - by Vineet
    I am running a Ubuntu Linux image in VMWare Player on my laptop. The host is Windows XP. There are two network adapters configured in VMWare Player - one is host-only and the other is bridged. Everything works fine as long as the host does not lose network connectivity. If the host loses connectivity (even briefly) and then re-acquires it, my bridged adapter in VMWare Player is still unable to get an IP address. The host-only adapter remains unaffected. Disconnecting/Reconnecting the bridged adapter does not help. Simply restarting the guest OS does not help. Restarting VMWare Player does not help. The only remedy is to reboot the host and then bring up VMWare Player all over again. Is there something I can try to avoid this reboot? I searched the existing questions but they seemed to talk about adapters in NAT mode, whereas I am interested in retaining Bridged mode.

    Read the article

< Previous Page | 270 271 272 273 274 275 276 277 278 279 280 281  | Next Page >