Search Results

Search found 25408 results on 1017 pages for 'back'.

Page 177/1017 | < Previous Page | 173 174 175 176 177 178 179 180 181 182 183 184  | Next Page >

  • MySQL Connect in Only 5 Days – Some Fun Stuff!

    - by Bertrand Matthelié
    72 1024x768 Normal 0 false false false EN-US X-NONE X-NONE /* Style Definitions */ table.MsoNormalTable {mso-style-name:"Table Normal"; mso-tstyle-rowband-size:0; mso-tstyle-colband-size:0; mso-style-noshow:yes; mso-style-priority:99; mso-style-qformat:yes; mso-style-parent:""; mso-padding-alt:0cm 5.4pt 0cm 5.4pt; mso-para-margin:0cm; mso-para-margin-bottom:.0001pt; mso-pagination:widow-orphan; font-size:11.0pt; font-family:"Calibri","sans-serif"; mso-ascii-font-family:Calibri; mso-ascii-theme-font:minor-latin; mso-fareast-font-family:"Times New Roman"; mso-fareast-theme-font:minor-fareast; mso-hansi-font-family:Calibri; mso-hansi-theme-font:minor-latin; mso-bidi-font-family:"Times New Roman"; mso-bidi-theme-font:minor-bidi;} We’ve recently blogged about the various MySQL Connect sessions focused on MySQL Cluster, InnoDB, the MySQL Optimizer and MySQL Replication. But we also wanted to draw your attention to some great opportunities to network and have fun! That’s also part of what makes a good conference... MySQL Connect Reception San Francisco Hilton - Continental Ballroom 6:30 p.m.–8:30 p.m. A great opportunity to network with Oracle’s MySQL engineers, partners having a booth in the exhibition hall and just about everyone at MySQL Connect. Long time MySQL users will see many familiar faces, and new users will be able to build valuable relationships. A must attend reception for sure! Taylor Street Open House 7:00 p.m.–9:00 p.m. After two intense days at MySQL Connect, you’ll get the chance to relax and continue networking at the Taylor Street Café Open House on Sunday evening. Perhaps recharging batteries for a full week at Oracle OpenWorld… The Oracle OpenWorld Music Festival Starting on Sunday eve and running through the entire duration of Oracle OpenWorld, the first Oracle OpenWorld Musical Festival features some of today’s breakthrough musicians. It’s five nights of back-to-back performances in the heart of San Francisco. Registered Oracle conference attendees get free admission, so remember your badge when you head to a show. More information here. You can check out the full MySQL Connect program here as well as in the September edition of the MySQL newsletter. Not registered yet? You can still save US$ 300 over the on-site fee – Register Now!

    Read the article

  • unity-panel apps running in the background

    - by blade19899
    i noticed that some of my apps are in the background in unity-panel. the apps are vmware-workstation,jupiter, empathy this is what i mean as you can see on the left some of my apps are running but in the background if i didn't set my unity-panel to see-through i wouldn't have noticed it. if i restart my laptop and start empathy it doesn't go in the background and also when i run compiz --replace the apps are back in the foreground this happens a lot when i am working my apps are suddenly in the unity-panel background?

    Read the article

  • How to handle interruptions in developer work without losing concentration? [closed]

    - by tomaszs
    I work as a developer for some years now. Mainly the issue why it's antisocial work is because you need to spend much time programming. I've been always the kind of developer who likes to cut off from any sources of distraction and spend several hours on project because in this way i (as i hope) do it faster. There are also other kinds of developers, more social that can chat, read, watch movies while development and they are ok with this and don't hesitate to be interrupted in their work in any time and come back to the project without any problem. For me any distraction is source of frustration because i need to spend substantial time to load my mind with all info about the project and to concentrate back on the tasks. I always thought it's better to do this that way because project is completed faster. But it makes some things difficult: it's hard to chat with someone who needs to have some important info: because you are a bit frustrated when you know you loose your Zen. And sometimes its more important to chat with someone than to loose Zen. Well.. mostly in any other kind of work the ability to be "multitask" is very important. But as a developer and as a person it's also very important to stay social. And i see now that the problem of concentration makes it difficult to make the right chose: the cost of maintaining concentration is just sometimes so damn high! So is it only me that i have so little concentration skills so any interruption is for me a big deal? Maybe it's just i have so bad memory so that i dont remember all issues of a project so long? Or maybe i develop the project in a fashion that requires me to store so much info on my mind only to be able to start working with code? Or should i just accept that being more social will make me finish project slower and in the fashion that i personally consider non 100% productive? And it's just normal thing and i should just accept it and start to live like any other person who has many works and don't assume that programming is in any case other than any other work and i just do fuzz about the whole concentration thing? This is question for mid-pro developers. I think you was having the same dillema in your life. I would be glad if you could help me take the right road here because it's just driving me and i suppose people i work with crazy for years.

    Read the article

  • Are Java servers really more preferable for web-development? [closed]

    - by Gerald Goward
    Many experienced people I know tell me that many web-projects, including enterprise ones, are better to develop with Java being back-end. Reasons: Ubuntu servers being cheaper and more reliable. MySql being much more "light" rather than "heavyweighted" MS Sql. I heard some more, but I really cant remember all of them. My question: I believe ASP.NET and Java are BOTH good for web-development and its just holywar subject. Am I right or not?

    Read the article

  • Ubuntu hangs on booting up after a update

    - by alFReD NSH
    I've made a clean install yesterday, for the first time restarted, everything went good and then after I updated packages and copied my old home directory to replace the new one, when I restarted it hung when it was booting. I tried reinstalling again and doing the same thing, but again same thing happened. Here's what I see, before when the Ubuntu logo with the five dots is shown: Then after that, 3 or 4 of the dots will load and hangs there. If I press arrow up before that, this will be shown I started my laptop again today(the pictures are for the day before) and after that, boot up with live CD and got the logs. dmesg: http://pastebin.com/aVxV7BQF syslog: http://pastebin.com/4E2BrRUK And some info: alfred@alFitop:~$ uname -a Linux alFitop 3.2.0-24-generic #39-Ubuntu SMP Mon May 21 16:52:17 UTC 2012 x86_64 x86_64 x86_64 GNU/Linux lshw: http://pastebin.com/AZbKJmsT sources.list : http://pastebin.com/2HazmuyV My problem is a bit similar to here: http://ubuntuforums.org/showthread.php?t=1918271 Though I didn't change my x.org config. Only changed home directory and updated packages. I've tried memtest and fschk, both passed. In the recovery mode boot option, I've also realized that same things happen in failsafe graphical mode. But when I go into the network mode, I can boot up my system, but of course same the graphics are just basic. Adding blacklist intel_ips to /etc/modprobe.d/blacklist.conf solves the first message, but still I get the broken pipe and CPU stack traces. The current kernel version is 3.2.0-25, I've tried booting up in the 3.2.0-23(the one the installer came with, but same results.) Also uninstalled apparmor, didn't help. I've installed Ubuntu again, this time without copying the home directory, also same result. --- UPDATE --- This problem was solved before with removing backports, but its back again! I've updated my laptop last night and the problem came back. It's definitely one of these packages. My /var/log/apt/term.log and /var/log/apt/history.log. I'm almost having the same situation. --- UPDATE --- I realized this also have happened on times that I have updated(haven't restarted after it) and my computer power has been cut off and its shutdown due to lack of power. And I realized if I just do as I answered but not in somewhere without GUI(networking mode has the GUI) it wouldn't work.

    Read the article

  • Picasa 3.9 login fails with 2-factor authentication

    - by Paul Pomes
    I've installed Picasa 3.9 via the instructions at webupd8, however the login window keeps failing with the message, "You must be connected to the Internet to use this feature." If "Try again" is tried I'll successfully pass the first login screen of username and password. Next I'm prompted for the verification code which then takes me back to the "You must be connected to the Internet to use this feature" screen again.

    Read the article

  • Lettre aux Payeurs de Novembre 2011/Global Payroll France November 2011 Newsletter

    - by Carolyn Cozart
    Global Payroll France November 2011 NewsletterPer the new publishing process started back in September, the Global Payroll France Newsletter is published on My Oracle Support. You will find the November 2011 letter under the following Doc ID: 1377174.1. Enjoy, Lettre aux Payeurs de Novembre 2011La Lettre aux payeurs française de novembre est disponible sur My Oracle Support, sous le numéro de document : 1377174.1, selon le nouveau processus de diffusion initié en septembre dernier. Bonne lecture,

    Read the article

  • Book Review: Fast Track to MDX

    - by Greg Low
    Another book that I re-read while travelling last week was Fast Track to MDX . I still think that it's the best book that I've seen for introducing the core concepts of MDX. SolidQ colleague Mark Whitehorn, along with Mosha Pasumansky and Robert Zare do an amazing job of building MDX knowledge throughout the book. I had dinner with Mark in London a few years back and I was pestering him to update this book. The biggest limitation of the book is that it was written for SQL Server 2000 Analysis Services,...(read more)

    Read the article

  • How to Do Competition Analysis

    One of the most important aspects of SEO is the work you put in before you even touch the website or build a single back link. This analysis work involves keyword research and competition analysis. Choose the wrong keywords and you could be wasting all your efforts in the onsite and offsite optimization. Choose keywords which have too much competition and you'll be taking on an uphill battle.

    Read the article

  • Search Engine Optimization - Article Branding Will Brand Your Business

    Branding your articles with your signature (a brief bio with each article) is the same as branding your business. This process helps to reach both goals in one shot. If you desire targeted traffic from the very start then give article marketing a serious shot and some serious effort. This form of marketing will also gain you multiple back links to your website of business and this is where your ranks start to increase.

    Read the article

  • Given the choice 8 out of 10 Optimisers prefer.........

    - by GrumpyOldDBA
    Did you know that included columns do not partake in the uniqueness of a unique index? ( see below ) A few months ago we upgraded our major production system from SQL2000 to SQL2008, this has allowed me to apply some of the index tuning techniques I devised for SQL2005 way back when to the current environment now we're confident we have no unexpected surprises to surface. Amongst the techniques I use is to pull information from the dmvs to find tables ( and indexes ) which are getting high numbers...(read more)

    Read the article

  • Right click is disabled on Desktop

    - by RameshKatkam
    I installed elementary OS and e17 on my Ubuntu 12.04 from ppas. After logout and login, right click on desktop is not working in gnome and unity but working in file manger.When I login into Pantheon session, I don't find an option to open Home folder and my login display manager theme is also changed.I want to get back to default Ubuntu login theme.Can i change the login theme with simple-lightdm manager. I dont want to mess up my system.Please hemp me how to fix these.

    Read the article

  • Blogging from the PASS Summit : Nov. 8th keynote

    - by AaronBertrand
    Douglas McDowell talks about day 1, the video montage featuring folks here from all over the world, and the fiscal year. The important point I took from this is that PASS is a non-profit committed to investing its revenue back into the community. They are hiring another full-time community evangelist, adding IT resources for online resources like the SQL Saturday site, and further expanding global efforts. He introduces the new board members: Wendy Pastrick, James Rowland-Jones, and Sri Sridharan....(read more)

    Read the article

  • dbus signal for volume up & down

    - by jldupont
    I recently upgrade to Ubuntu 10.10 Maverick Meerkat and to my dismay, the mediakeys "volume up" and "volume down" do not send Dbus signals anymore... how can I add these back? Thanks!! Update: it seems that under some circumstances (which I don't know exactly yet), the DBus signals start working again. It is as though when a certain application (TBD) is executed, the dbus signals are re-activated.

    Read the article

  • Pay in the future should make you think in the present

    - by BuckWoody
    Distributed Computing - and more importantly “-as-a-Service” models of computing have a different cost model. This is something that sounds obvious on the surface but it’s often forgotten during the design and coding phase of a project. In on-premises computing, we’re used to purchasing a server and all of the hardware infrastructure and software licenses needed not only for one project, but several. This is an up-front or “sunk” cost that we consume by running code the organization needs to perform its function. Using a direct connection over wires you’ve already paid for, we don’t often have to think about bandwidth, hits on the data store or the amount of compute we use - we just know more is better. In a pay-as-you-go model, however, each of these architecture decisions has a potential cost impact. The amount of data you store, the number of times you access it, and the amount you send back all come with a charge. The offset is that you don’t buy anything at all up-front, so that sunk cost is freed up. And financial professionals know that money now is worth more than money later. Saving that up-front cost allows you to invest it in other things. It’s not just that you’re using things that now cost money - it’s that the design itself in distributed computing has a cost impact. That can be a really good thing, such as when you dynamically add capacity for paying customers. If you can tie back the cost of a series of clicks to what a user will pay to do so, you can set a profit margin that is easy to track. Here’s a case in point: Assume you are using a large instance in Windows Azure to compute some data that you retrieve from a SQL Azure database. If you don’t monitor the path of the application, you may not know what you are really using. Since you’re paying by the size of the instance, it’s best to maximize it all the time. Recently I evaluated just this situation, and found that downsizing the instance and adding another one where needed, adding a caching function to the application, moving part of the data into Windows Azure tables not only increased the speed of the application, but reduced the cost and more closely tied the cost to the profit. The key is this: from the very outset - the design - make sure you include metrics to measure for the cost/performance (sometimes these are the same) for your application. Windows Azure opens up awesome new ways of doing things, so make sure you study distributed systems architecture before you try and force in the application design you have on premises into your new application structure.

    Read the article

  • What is happening to a domain with status PENDING DELETE as well as AUTORENEWPERIOD?

    - by Alex Angas
    A domain I once registered but gave away: Expiration Date:13-Mar-2013 14:39:45 UTC Sponsoring Registrar:Directi Internet Solutions Pvt. Ltd. dba PublicDomainRegistry.com (R159-LRMS) Status:PENDING DELETE RESTORABLE Status:HOLD Status:AUTORENEWPERIOD Status:REDEMPTIONPERIOD Registrant ID:DI_7838158 Registrant Name:scherhag Registrant Organization:Locafroid europe sa I'd like to get this domain back and I'm hoping that PENDING DELETE means it might soon be released to the market. However I'm not sure, as there is also a status that says AUTORENEWPERIOD?

    Read the article

  • .Net Micro

    - by MarkPearl
    A while back I purchased a RFID scanner that could be connected to a PC and programmed via VS. It was a fun purchase an though the import duties nailed me, I was glad to get the little gadget. Last night while listening to .Net Rocks I heard of another company that sells similar components for .Net Micro. Check out their websites…. TinyClr GHI Electronics .Net Micro Website Trossen Robotics

    Read the article

  • User Group Meeting in Dundee (Scotland)

    Tony Rogerson and myself are going to be in Dundee (we are going back to school) week beginning 18.1.2010.  Whilst there we are going to do a User Group meeting and it would be great to see people there.  More details can be found here - http://sqlserverfaq.com?eid=211

    Read the article

  • Looking for new language and new technology [closed]

    - by Basim
    back when Microsoft relased .Net in 2002 or whatever, when I look at that time I say to myself what I if I picked one of Microsoft language in that time and still work on it, of course I will be professional by now. I am looking for a new language that is going up and will be big thing in the next 5-10 years, so in that time i can see the big picture and I know that I'm one of the few people who started from the beginning with X programming language or technology. My interest is web development.

    Read the article

  • DBA Best Practices - A Blog Series: Episode 1 - Backups

    - by Argenis
      This blog post is part of the DBA Best Practices series, on which various topics of concern for daily database operations are discussed. Your feedback and comments are very much welcome, so please drop by the comments section and be sure to leave your thoughts on the subject. Morning Coffee When I was a DBA, the first thing I did when I sat down at my desk at work was checking that all backups have completed successfully. It really was more of a ritual, since I had a dual system in place to check for backup completion: 1) the scheduled agent jobs to back up the databases were set to alert the NOC in failure, and 2) I had a script run from a central server every so often to check for any backup failures. Why the redundancy, you might ask. Well, for one I was once bitten by the fact that database mail doesn't work 100% of the time. Potential causes for failure include issues on the SMTP box that relays your server email, firewall problems, DNS issues, etc. And so to be sure that my backups completed fine, I needed to rely on a mechanism other than having the servers do the taking - I needed to interrogate the servers and ask each one if an issue had occurred. This is why I had a script run every so often. Some of you might have monitoring tools in place like Microsoft System Center Operations Manager (SCOM) or similar 3rd party products that would track all these things for you. But at that moment, we had no resort but to write our own Powershell scripts to do it. Now it goes without saying that if you don't have backups in place, you might as well find another career. Your most sacred job as a DBA is to protect the data from a disaster, and only properly safeguarded backups can offer you peace of mind here. "But, we have a cluster...we don't need backups" Sadly I've heard this line more than I would have liked to. You need to understand that a cluster is comprised of shared storage, and that is precisely your single point of failure. A cluster will protect you from an issue at the Operating System level, and also under an outage of any SQL-related service or dependent devices. But it will most definitely NOT protect you against corruption, nor will it protect you against somebody deleting data from a table - accidentally or otherwise. Backup, fine. How often do I take a backup? The answer to this is something you will hear frequently when working with databases: it depends. What does it depend on? For one, you need to understand how much data your business is willing to lose. This is what's called Recovery Point Objective, or RPO. If you don't know how much data your business is willing to lose, you need to have an honest and realistic conversation about data loss expectations with your customers, internal or external. From my experience, their first answer to the question "how much data loss can you withstand?" will be "zero". In that case, you will need to explain how zero data loss is very difficult and very costly to achieve, even in today's computing environments. Do you want to go ahead and take full backups of all your databases every hour, or even every day? Probably not, because of the impact that taking a full backup can have on a system. That's what differential and transaction log backups are for. Have I answered the question of how often to take a backup? No, and I did that on purpose. You need to think about how much time you have to recover from any event that requires you to restore your databases. This is what's called Recovery Time Objective. Again, if you go ask your customer how long of an outage they can withstand, at first you will get a completely unrealistic number - and that will be your starting point for discussing a solution that is cost effective. The point that I'm trying to get across is that you need to have a plan. This plan needs to be practiced, and tested. Like a football playbook, you need to rehearse the moves you'll perform when the time comes. How often is up to you, and the objective is that you feel better about yourself and the steps you need to follow when emergency strikes. A backup is nothing more than an untested restore Backups are files. Files are prone to corruption. Put those two together and realize how you feel about those backups sitting on that network drive. When was the last time you restored any of those? Restoring your backups on another box - that, by the way, doesn't have to match the specs of your production server - will give you two things: 1) peace of mind, because now you know that your backups are good and 2) a place to offload your consistency checks with DBCC CHECKDB or any of the other DBCC commands like CHECKTABLE or CHECKCATALOG. This is a great strategy for VLDBs that cannot withstand the additional load created by the consistency checks. If you choose to offload your consistency checks to another server though, be sure to run DBCC CHECKDB WITH PHYSICALONLY on the production server, and if you're using SQL Server 2008 R2 SP1 CU4 and above, be sure to enable traceflags 2562 and/or 2549, which will speed up the PHYSICALONLY checks further - you can read more about this enhancement here. Back to the "How Often" question for a second. If you have the disk, and the network latency, and the system resources to do so, why not backup the transaction log often? As in, every 5 minutes, or even less than that? There's not much downside to doing it, as you will have to clear the log with a backup sooner than later, lest you risk running out space on your tlog, or even your drive. The one drawback to this approach is that you will have more files to deal with at restore time, and processing each file will add a bit of extra time to the entire process. But it might be worth that time knowing that you minimized the amount of data lost. Again, test your plan to make sure that it matches your particular needs. Where to back up to? Network share? Locally? SAN volume? This is another topic where everybody has a favorite choice. So, I'll stick to mentioning what I like to do and what I consider to be the best practice in this regard. I like to backup to a SAN volume, i.e., a drive that actually lives in the SAN, and can be easily attached to another server in a pinch, saving you valuable time - you wouldn't need to restore files on the network (slow) or pull out drives out a dead server (been there, done that, it’s also slow!). The key is to have a copy of those backup files made quickly, and, if at all possible, to a remote target on a different datacenter - or even the cloud. There are plenty of solutions out there that can help you put such a solution together. That right there is the first step towards a practical Disaster Recovery plan. But there's much more to DR, and that's material for a different blog post in this series.

    Read the article

  • Elantech multitouch breaks after sleep and unknown events

    - by levesque
    EDIT: I previously thought this was related to mouse, but I have been experiencing these issues without a mouse plugged-in during the past weeks. I therefore rewrote the description. The two-finger scrolling on my UX31 works just fine, but sometimes it stops working, the cause is unknown. Multitouch also breaks whenever I go in sleep mode. Logging out and back in fixes the issue, till it happens again (sleep, or the unknown source) Does this sound familiar to anyone here?

    Read the article

  • After locking the screen in Ubuntu 14.04, password is not accepted, How can it be fixed?

    - by Itai Ganot
    I'm running Ubuntu 14.04 fully updated on my laptop. Since the last update every time I lock the screen (when leaving my room for example) - when I get back and input my password, it is not accepted even though it's the correct password, the error I get is: Password incorrect, please try again I found that clicking the "Switch Account" fixes the issue but it is very annoying, if you know any way to fix it, it would be nice. Thanks in advance

    Read the article

  • Custom errors won't turn off (2 replies)

    ..NET Framework 3.5 Visual Studio 2008 C# I implemented my own transport channel. It works to a point: the client is capable of invoking a method on a server singleton. When the method completes successfully it exits and back on the client I receive: System.Runtime.Remoting.RemotingException was unhandled Message &quot;Server encountered an internal error. For more information, turn off customErrors in...

    Read the article

  • Linkedin recommendations, useful? [closed]

    - by scottyab
    Linkedin has been around a few years now and while not everyone is keen on it. Chatting to a few people they viewed Linkedin recommendations as mutual back starching where poor candidates can recommend each other and and provide little value. I take a honest approach when recommending colleagues, trying to note specifics that person has performed well at and don't recommend colleagues on linkedin who wouldn't recommend in real life. What regard do people hold the Linkedin recommendations? would it effect your hiring decision?

    Read the article

  • Ugly right-click menu and menubar in some applications

    - by halflings
    After installing some packages (unfortunately I can't remember which ones) and installing some GTK3 themes (Faience / Faenza), my terminal's menubar and the right-click menu (in some contexts) started to look very ugly ! 1: Right-click menu 2: Terminal menubar, notice the white rounded corner rectangles, it looks WAY uglier on other themes Note that the menubar isn't "ugly" in most contexts, but it still is on the desktops and some applications. (For the menubar, I think the terminal is the only one concerned by this) I tried going back to my original theme, it didn't fix the problem.

    Read the article

< Previous Page | 173 174 175 176 177 178 179 180 181 182 183 184  | Next Page >