Search Results

Search found 24201 results on 969 pages for 'andrew case'.

Page 472/969 | < Previous Page | 468 469 470 471 472 473 474 475 476 477 478 479  | Next Page >

  • use network drives as mount points during installation?

    - by ajsie
    is it possible to use network storage locations as mount points during installation? cause i want to separate system (ubuntu) with data (personal files). eg. if i have 5 computers i don't want to recreate /home/david 5 times. so i want to mount networkdrive/home to /home in local ubuntu server. so ALL users home folders could be used and maybe also networkdrive/projects to /projects. in that way its ok if i by accident repartitioned the local ubuntu server cause all data is not there on that server, but in the data server. is separating "data" from "logic" good in this case? and is it possible? what protocol should i use for the mapping over internet? (maybe the server is in Sweden, and the data is in Norway). thanks.

    Read the article

  • Macbook Pro (SantaRosa) internal display not detected by graphics card, external monitor OK

    - by BLAU
    My MacBook Pro (2.4 Ghz, Santa Rosa with infamous nVidia card) acts strange. It shows the normal gray screen with Apple logo and animation flawlessly during start up but the internal display goes black without any rendering at all when all is loaded. (shining a light on display show nothing) If an external monitor is connected through the DVI port it will remain black during start up and then show the desktop as the internal display goes black. This happens both while booting to Mountain Lion and Windows XP. I have checked "About my Mac" and only the external display is listed. The same is the case if I use the nVidia Control panel in Windows XP. My questions: Is this a hardware problem or is it related to software maybe even firmware? What controls the display during start up, graphics card or something else?

    Read the article

  • Longlasting Macport installation

    - by user10826
    Hi, I use Macports on Mac OSX to install some software. But there are somethings that I find very strange. For instance, yesterday I installed some app. and the it needed to get gcc43 and compile completely from source! It took a lot of time, although now everything works right. I understand that compilation of everything in the gentoo spirit is eomthing nice, but in the Mac case, not necessary. There are ony sole hardware variations, so it will be easier to get the binaries. I wonder if Macporst can be forced to get the binaries when available and therefore avoid these time consuming compilations. Thanks

    Read the article

  • Domain Transfer Protection - need advice

    - by Jack
    Hey, I am about to purchase a domain name for a bit of money. I do not personally know the person who I am purchasing the domain name from, we have only chatted via email. The proposed process for the transfer is: The owner of the domain lowest the domain name security and emails me the domain password, I request the transfer After the request, I transfer the money via PayPal When the money has been cleared the current domain name owner confirms the transfer via the link that he receives in that email I wait for it to be transferred. The domain is currently registered with DirectNIC - http://www.directnic.com/ Is this the best practice? Seeing I am paying a bit of money for this domain name, I am worried that after the money has been cleared that I won't see the domain name or hear from the current domain name owner again. Is there a 'domain governing body' which I can report to if this is the case? Is the proposed transfer process the best solution? Any advice would be awesome. Thanks! Jack

    Read the article

  • What to do when TDD tests reveal new functionality that is needed that also needs tests?

    - by Joshua Harris
    What do you do when you are writing a test and you get to the point where you need to make the test pass and you realize that you need an additional piece of functionality that should be separated into its own function? That new function needs to be tested as well, but the TDD cycle says to Make a test fail, make it pass then refactor. If I am on the step where I am trying to make my test pass I'm not supposed to go off and start another failing test to test the new functionality that I need to implement. For example, I am writing a point class that has a function WillCollideWith(LineSegment): public class Point { // Point data and constructor ... public bool CollidesWithLine(LineSegment lineSegment) { Vector PointEndOfMovement = new Vector(Position.X + Velocity.X, Position.Y + Velocity.Y); LineSegment pointPath = new LineSegment(Position, PointEndOfMovement); if (lineSegment.Intersects(pointPath)) return true; return false; } } I was writing a test for CollidesWithLine when I realized that I would need a LineSegment.Intersects(LineSegment) function. But, should I just stop what I am doing on my test cycle to go create this new functionality? That seems to break the "Red, Green, Refactor" principle. Should I just write the code that detects that lineSegments Intersect inside of the CollidesWithLine function and refactor it after it is working? That would work in this case since I can access the data from LineSegment, but what about in cases where that kind of data is private?

    Read the article

  • Congratulations to 2012 Innovation Award winners in BPM category

    - by Manoj Das
    Last year many of our customers went live on BPM 11g. It is my extreme pleasure to congratulate two of them – Amadeus and Navistar – for being awarded Oracle Fusion Middleware Innovation Award at Oracle OpenWorld 2012. We invited our customers to submit their most innovative BPM implementations that have delivered substantiated value to them. This year we saw more than 20 submissions from our customers seeing significant business value from their live BPM 11g deployments. The submissions came from across the world, spanning various industry verticals including manufacturing, healthcare, logistics, Hi-Tech, Public Sector, Education and covering many process usage patterns. Award submissions were evaluated based on the uniqueness of their business case, business benefits, level of impact relative to the size of the organization, complexity and magnitude of implementation, and the originality of architecture. Amadeus Team Receiving Innovation Award from Hasan Rizvi Congratulations to Amadeus and Navistar and their teams on being recognized from among some very strong submissions and more importantly for the business value delivered. It is an honor to be part of your success and to play a small role in the innovation you drive. Navistar is a leading truck manufacturing company which produces International® brand commercial and military trucks, MaxxForce® brand diesel engines, IC Bus™ brand school and commercial buses, and Navistar RV brands of recreational vehicles. The company also provides truck and diesel engine service parts. Amadeus is a leading transaction processor for the global travel and tourism industry, providing transaction processing power and technology solutions to both travellers and travel providers. Both Navistar and Amadeus have leveraged Oracle BPM Suite to improve visibility into their business and made their business more agile and efficient. We congratulate them again and wish them continued success in their business and future BPM initiatives.

    Read the article

  • Online Windows Server Virtual Hard Drive Images for Use with VMWare Server 1.0

    - by charfeddine.ahmed
    Hello there, I have a remote Server running the VMWare hypervisor. I want to create a virtual machine running Windows Server. However I can't upload mine since my internet connection is slow (would take me days to do the upload). Microsoft has public hard drive images with Windows Server Trial on them, but they work for Virtual PC. I am looking for such files that can be used with VMWare Server. That case I can download these files directly to the server which enjoys a fast connection. Thanks in advance.

    Read the article

  • fdisk (linux) partitioning raid0

    - by silverrocker
    'm trying to create partitions for a slackware instalation on my computer (beside Windows 7) just to have a nice distro running mostly for school but when I run fdisk and print the partitiontable I get the following message: Partition x does not end on cylinder boundary. (in my case x = 1, just using x to help googlers). I must say I'm using a raid card (AMCC 3ware 9500S SATA RAID Controller). Maybe this is the problem. How can I fix this without loosing any data? (I've posted this qeustion on stack overflow aswell but then people sugjested it should be moved to here, but I don't know how to move my question so I just posted it here manually (I hope that's ok))

    Read the article

  • OWA draft folder doesn't sync up wint Outlook 2007 (when using Citrix)

    - by George
    I have a user logging in to Citrix Server (on Windows 2003) to use Outlook 2007. In OWA, he sees all his drafts in Draft Folder and can easily access them, but when he is Citrix, he can see the folder, but not the messages. I had him check Normal and Favorite Folders under View - Navigation Pane as well as execute outlook /cleanviews to no help. I should also clarify, we host exchange locally and it syncs up with Outlook 2007 in Citrix. Remote users use either OWA for access or login to Citrix and use Outlook 2007. In his case ALL folders appear in Outlook 2007, but draft folder doesn't show any saved messages, even though in OWA messages are there and he can edit, delete and send them. Please, help! Thanks!!!

    Read the article

  • Google is not treating two Australian schools as separate sites when both are subdomains of qld.edu.au

    - by LuckySpoon
    My question relates to two websites, each of which is a "Calvary Christian College", however in two totally different locations and unrelated to each other entirely (except by name, and thus domain). All schools in the state are issued a <school-name>.qld.edu.au subdomain, in this case calvary.qld.edu.au and calvarycc.qld.edu.au. Now what's interesting is that these domains are crossing each other in sitelinks for searches such as calvary christian college townsville. The green data here is for one school (the Townsville school, as per search term), and the red data is for the other school. I've put a demotion in for this 6 months ago (we control calvary.qld.edu.au), however we're seeing no change on the results page. I have been able to get the owners of calvarycc.qld.edu.au to submit demotions for our domain, which should go in sometime in the next few days. What can we do to tell Google that these websites are not interchangeable, despite both appearing as "subdomains" of qld.edu.au? We can possibly open channels of communication with the administrators of qld.edu.au but will need to tell them what we need to change, and at this point I'm out of ideas.

    Read the article

  • Encryption over gigabit carrier ethernet

    - by Roy
    I would like to encrypt traffic between two data centres. Communication between the sites is provided as a standard provider bridge (s-vlan/802.1ad), so that our local vlan tags (c-vlan/802.1q) are preserved on the trunk. The communication traverse several layer 2 hops in the provider network. Border switches on both sides are Catalyst 3750-X with the MACSec service module, but I assume MACSec is out of the question, as I don't see any way to ensure L2 equality between the switches over a trunk, although it may be possible over a provider bridge. MPLS (using EoMPLS) would certainly allow this option, but is not available in this case. Either way, equipment can always be replaced to accommodate technology and topology choices. How do I go about finding viable technology options that can provide layer 2 point-to-point encryption over ethernet carrier networks?

    Read the article

  • Recommend a good DVD or BD replacement drive that can read older CD-R/DVD+R media?

    - by Irinotecan
    So I have a bunch of older CD-R (and a few DVD+R) discs that are either suffering from "bit rot", or a case of crappy no-name Chinese DVD drives being unable to read some or all of anymore. I just threw my last no-name DVD drive in the trash after it ejected a disk still spinning causing it to scratch to the point of being useless, so I'm looking for a replacement drive this time around with an eye for quality over a dirt cheap price. I'd prefer a BD drive, but I'll take any suggestions for good, reliable DVD or BD drives that people have reported having good success with for reading older burned media. Any takers?

    Read the article

  • Mirroring MySQL server with diffrent configuration

    - by HTF
    I have to migrate MySQL server to a different data centre so I would like to create another MySQL slave server in new DC and then promote it to a master later on. I previously used LVM snapshots and Percona Xtrabackup for this purpose but this time I've optimized MySQL configuration file that prevents me from using these methods. Old server (backup): innodb_log_file_size = 256M innodb_log_files_in_group = 3 New server (restore): innodb_log_file_size = 512M innodb_log_files_in_group = 2 The Xtrabackup script and LVM snapshots copy the whole directory structure so the MySQL server won't start because there is a different size for InnoDB logs. Is there any solution to avoid a downtime in this case? I can't use mysqldumps as there is around 8000 databases so I would have to take the server down for a couple of hours. I was also thinking to use the old settings with Xtrabackup and then change it once the new server is promoted to a master - less downtime but I'm not sure if this will work? Thank you Regards

    Read the article

  • FCoE, on any Ethernet switch?

    - by javano
    I understand the concept of FCoE. I have looked at the Wikipedia page and looking at the layer 2 frame diagram, it looks like FCoE really should "just work" on any Ethernet switch, but is this really the case? If so, what do switches like Cisco's Nexus 5k or 6120P offer that normal switches don't (in specific relation to FCoE)? I am just using those two switches as examples. On the Nexus 5548UP page for example it says the following; Unified ports that support traditional Ethernet, Fibre Channel (FC),and Fibre Channel over Ethernet (FCoE) Well if FCoE runnins over regular Ethernet, that why does it support "Ethernet and Fibire Channel over Ethernet"? This is why I am curious as to weather FCoE will run on any Ethernet switch and these switches just support "bonus" features, or if you do indeed require a specialist switch. Thank you.

    Read the article

  • Asking for Credentials, when requesting shared folders on the server, regularly for domain users?

    - by MFH
    In our network, single domain controller, when some users (members of the domain) request shared folders on the server they are required to enter their credentials, the server is Windows Server 2008 R2, even checking to remember my credentials doesn't work, sometimes it shows this message: "The system has detected a possible attempt to compromise security. Please ensure that you can contact the server that authenticated you", sometimes it shows different messages, when I try to recreate the case sometimes I failed, I searched Google for it, I didn't find useful results, some talk about kerberos, but we don't use kerberos, this keeps going each day or two, how to overcome this, I don't want these messages to appear to users?

    Read the article

  • Merging Two Git Repositories with branches

    - by Joel K
    I realize there's a Stack Overflow question: http://stackoverflow.com/questions/277029/combining-multiple-git-repositories But I haven't found git-stitch-repo to be quite the tool I'm looking for. I also consider this more of a sysadmin task. How do I take code from an external repository and combine it with code from a primary repository while maintaining history/diffs and branches. Use case: An outside development team using SVN has ported to git and now wants to 'merge' their code in to the main company's git repo. I've tried subtree merges, but I lose the history. I've tried git-stitch-repo, but that process results in an entirely new repo that's missing branches. I just want to slot in some outside code as a sub-directory in our current main repo with as little disruption as possible and while maintaining the other project's history. Any success stories out there?

    Read the article

  • How should I track approval workflow when users at every security level can create a request?

    - by Eric Belair
    I am writing a new application that allows users to enter requests. Once a request is entered, it must follow an approval workflow to be finally approved by a user the highest security level. So, let's say a user at Security Level 1 enters a request. This request must be approved by his superior - a user at Security Level 2. Once the Security Level 2 user approves it, it must be approved by a user at Security Level 3. Once the Security Level 3 user approves it, it is considered fully approved. However, users at any of the three Security Levels can enter requests. So, if a Security Level 3 user enters a request, it is automatically considered "fully approved". And, if a Security Level 2 user enters a request, it must only be approved by a Security Level 3 user. I'm currently storing each approval status in a Database Log Table, like so: STATUS_ID (PK) REQUEST_ID STATUS STATUS_DATE -------------- ------------- ---------------- ----------------------- 1 1 USER_SUBMIT 2012-09-01 00:00:00.000 2 1 APPROVED_LEVEL2 2012-09-01 01:00:00.000 3 1 APPROVED_LEVEL3 2012-09-01 02:00:00.000 4 2 USER_SUBMIT 2012-09-01 02:30:00.000 5 2 APPROVED_LEVEL2 2012-09-01 02:45:00.000 My question is, which is a better design: Record all three statuses for every request ...or... Record only the statuses needed according to the Security Level of the user submitting the request In Case 2, the data might look like this for two requests - one submitted by Security Level 2 User and another submitted by Security Level 3 user: STATUS_ID (PK) REQUEST_ID STATUS STATUS_DATE -------------- ------------- ---------------- ----------------------- 1 3 APPROVED_LEVEL2 2012-09-01 01:00:00.000 2 3 APPROVED_LEVEL3 2012-09-01 02:00:00.000 3 4 APPROVED_LEVEL3 2012-09-01 02:00:00.000

    Read the article

  • Connect to NFS on availability

    - by berkes
    What would be a good way to automatically mount an NFS when it gets/is available? I have the following: Media server at home, running Ubuntu, 10.10 with GUI *) Laptop often at home, often on the road or at clients. Ubuntu 10.10 with GUI. What I'd like is my laptop connecting to the nfs (or any other mountable networked filesystem) so that Banshee sees all the music, new podcast-entries (and video) from that media-server. I already have firefly (mt-daapd) running, which works, but is flakey on both server-side and client-side. But its biggest downside, is that I cannot easily fix metadata on files on the media-server this way. DAAP is read-only by design. I can mount nfs manually, through a sudo mount /media/nfsmultimedia/. I am not looking for a manual, or howto on setting up a NFS client and server. Merely a way to have this more transparently working. Obviously I'd like the NFS to be unmounted if the network is no longer available (i.e. when I open my laptop-lid on my clients buro). It may be, that an NFS is not suited for this, in that case, I'd love to hear other options. :) *) Actually: I also have a fileserver, backupserver and webserver to which I'd like to connect in a somewhat similar way. Right now I connect to these over SSH, using gvfs.

    Read the article

  • Should we test all our methods?

    - by Zenzen
    So today I had a talk with my teammate about unit testing. The whole thing started when he asked me "hey, where are the tests for that class, I see only one?". The whole class was a manager (or a service if you prefer to call it like that) and almost all the methods were simply delegating stuff to a DAO so it was similar to: SomeClass getSomething(parameters) { return myDao.findSomethingBySomething(parameters); } A kind of boilerplate with no logic (or at least I do not consider such simple delegation as logic) but a useful boilerplate in most cases (layer separation etc.). And we had a rather lengthy discussion whether or not I should unit test it (I think that it is worth mentioning that I did fully unit test the DAO). His main arguments being that it was not TDD (obviously) and that someone might want to see the test to check what this method does (I do not know how it could be more obvious) or that in the future someone might want to change the implementation and add new (or more like "any") logic to it (in which case I guess someone should simply test that logic). This made me think, though. Should we strive for the highest test coverage %? Or is it simply an art for art's sake then? I simply do not see any reason behind testing things like: getters and setters (unless they actually have some logic in them) "boilerplate" code Obviously a test for such a method (with mocks) would take me less than a minute but I guess that is still time wasted and a millisecond longer for every CI. Are there any rational/not "flammable" reasons to why one should test every single (or as many as he can) line of code?

    Read the article

  • Ctrl W s not splitting windows in vim

    - by rajan sthapit
    I am trying to use Ctrl+ws to split a window in vim. However it is not working in my case. I opened a file using vim filename. Then I pressed Ctrl+ws. But as soon as I press just Ctrl+w it clears the viewport and I could see my shell display before vim opened the file. I mean the view port is replaced by the content just before the file was opened. However, I am still editing the file with vim. Suggestions?

    Read the article

  • Developing momentum on open source projects

    - by sashang
    Hi I've been struggling to develop momentum contributing to open source projects. I have in the past tried with gcc and contributed a fix to libstdc++ but it was a once off and even though I spent months in my spare time on the dev mailing list and reading through things I just never seemed to develop any momentum with the code. Eventually I unsubscribed and got my free time back and uncluttered my mailbox. Like a lot of people I have some little open source defunct projects lying around on the net, but they're not large and I'm the only contributor. At the moment I'm more interested in contributing to a large open source project and want to know how people got started because I find it difficult while working full time to develop any momentum with the code base. Other more regular contributors, who are on the project full-time, are able to make changes at will and as result enter that positive feedback cycle where they understand the code and also know where it's heading. It makes the barrier to entry higher for those that come along later. My questions are to people who actively contribute to large opensource projects, like the Linux kernel, or gcc or clang/llvm or anything else with say a developer head count of more than 10. How did you get started? Was there a large chunk of time in your life that you just could dedicate to working on the project? I know in Linus's case he had a chunk of time (6 months) to get it started. What barriers to entry did you encounter? Can you describe the initial stages of the time spent with the project, from when you had little understanding of the code to when you understood enough to commit regularly. Thanks

    Read the article

  • Mixing JavaFX, HTML 5, and Bananas with the NetBeans Platform

    - by Geertjan
    The banana in the image below can be dragged. Whenever the banana is dropped, the current date is added to the viewer: What's interesting is that the banana, and the viewer that contains it, is defined in HTML 5, with the help of a JavaScript and CSS file. The HTML 5 file is embedded within the JavaFX browser, while the JavaFX browser is embedded within a NetBeans TopComponent class. The only really interesting thing is how drop events of the banana, which is defined within JavaScript, are communicated back into the Java class. Here's how, i.e., in the Java class, parse the HTML's DOM tree to locate the node of interest and then set a listener on it. (In this particular case, the event listener adds the current date to the InstanceContent which is in the Lookup.) Here's the crucial bit of code: WebView view = new WebView(); view.setMinSize(widthDouble, heightDouble); view.setPrefSize(widthDouble, heightDouble); final WebEngine webengine = view.getEngine(); URL url = getClass().getResource("home.html"); webengine.load(url.toExternalForm()); webengine.getLoadWorker().stateProperty().addListener( new ChangeListener() { @Override public void changed(ObservableValue ov, State oldState, State newState) { if (newState == State.SUCCEEDED) { Document document = (Document) webengine.executeScript("document"); EventTarget banana = (EventTarget) document.getElementById("banana"); banana.addEventListener("click", new MyEventListener(), true); } } }); It seems very weird to me that I need to specify "click" as a string. I actually wanted the drop event, but couldn't figure out what the arbitrary string was for that. Which is exactly why strings suck in this context. Many thanks to Martin Kavuma from the Technical University of Eindhoven, who I met today and who inspired me to go down this interesting trail.

    Read the article

  • Problem with missing JSON functions on PHP 5.2.6 / Plesk 8.4

    - by Drachenviech
    I have a vserver running openSuse 10.3, Apache 2 and Plesk 8.4. I can update/upgrade neither, as it is apparently not recommended to upgrade openSuse 10.3 (and an update to the EOL 10.4 does not seem to make much sense) and Plesk fails to update no matter what version I try (even fails to upgrade to 8.4.1). Still I can live with that somehow, primarily because I don’t have the time to do a fresh remote install on the vserver. What really is a problem is, that though the installed PHP is 5.2.6 it has no zip library and no json functions. The first is probably because PHP was not compiled with --enable-zip. The second is a big mystery though. As I understand it, it always comes with PHP unless its compiled with the --disable-json configure option. This is however not the case. And the json extension module is just not there. I even tried to enable it with extension=json.so with no luck either. the configure options of my PHP are (as shipped with Plesk 8.4) '../configure' '--prefix=/usr' '--datadir=/usr/share/php5' '--mandir=/usr/share/man' '--bindir=/usr/bin' '--with-libdir=lib' '--includedir=/usr/include' '--sysconfdir=/etc/php5/apache2' '--with-config-file-path=/etc/php5/apache2' '--with-config-file-scan-dir=/etc/php5/conf.d' '--enable-libxml' '--enable-session' '--with-mm' '--with-pcre-regex=/usr' '--enable-xml' '--enable-simplexml' '--enable-spl' '--enable-filter' '--disable-debug' '--enable-inline-optimization' '--disable-rpath' '--disable-static' '--enable-shared' '--program-suffix=5' '--with-pic' '--with-gnu-ld' '--with-system-tzdata=/usr/share/zoneinfo' '--with-apxs2=/usr/sbin/apxs2' '--disable-all' '--disable-cli' As I understand it, PECL is not an option with 5.2.6. Or am I mistaken? Even if I was not, the openSuse repository only goes as far as PHP 5.2.4. The openSuse install even came without zypper, which I had to manually install. So is there a way to get ziplib and json running in PHP 5.2.6 without having to recompile the binary?

    Read the article

  • Why is apt-cache so slow?

    - by Damn Terminal
    After upgrade to Trusty (14.04) from Saucy (13.10), all apt operations are very slow. Even those that do not include downloading anything, or connecting to any servers. For example, displaying the apt policy # time apt-cache policy [...] real 0m8.951s user 0m5.069s sys 0m3.861s takes almost ten seconds! Mostly a weird lag right after issuing the command. And it's the same even if I issue the same command again. On another system it doesn't take a tenth of a second real 0m0.096s user 0m0.070s sys 0m0.023s The other system is a little beefier but there was no noticeable difference before the upgrade. It's the same with apt-get, anything apt-related. How do I find out the source of this lag and fix it? Additional info: # cat /etc/nsswitch.conf # /etc/nsswitch.conf # # Example configuration of GNU Name Service Switch functionality. # If you have the `glibc-doc-reference' and `info' packages installed, try: # `info libc "Name Service Switch"' for information about this file. passwd: compat group: compat shadow: compat hosts: files dns networks: files protocols: db files services: db files ethers: db files rpc: db files netgroup: nis BTW is my understanding of how apt-cache works correct? It doesn't make any network connections when I run apt-cache policy, right? In case I'm wrong and it matters, here are my sources https://gist.github.com/anonymous/02920270ff68e23fc3ec

    Read the article

  • External display resolution issue

    - by Steven
    I'm new to Ubuntu. I've 'dabbled' in the past but Windows was annoying me too much so I finally made the change. One thing I use my laptop for is outputting the screen onto an external monitor. In Windows this usually works fine (but when there is an issue, it's a pain to resolve and has crashed my computer on many occasions). In Ubuntu it's a very simple case of "plug and play" because it displays the content immediately with no major problems. However, I can only alter the resolution of the external display to 4:3 formats, whereas the external monitor is 16:9, so the picture is stretched. In Windows there's loads of resolutions to choose from. If I disable the laptop's screen, it still doesn't allow me to display the correct resolution externally (on VGA). Further to that, if I close the laptop it keeps the external screen as it is, but when I reopen the laptop, the laptop screen doesn't turn back on until I restart. If it's a video I am watching, I can use VLC to alter the aspect ratio so the video looks fine, but it's quite an annoying work around. I used terminal to find my graphics controller, which is "Intel Corporation Mobile 4 Series Chipset Integrated Graphics Controller [8086:2a42] (rev 07)" (not great, I know... the laptop was a free gift with my mobile!) Any idea of how I can get the resolution on the external monitor to be correct?

    Read the article

< Previous Page | 468 469 470 471 472 473 474 475 476 477 478 479  | Next Page >